Is AI A Good Companion? INTIMA Tests This Comprehensively

đ usncan Note: Is AI A Good Companion? INTIMA Tests This Comprehensively
Disclaimer: This content has been prepared based on currently trending topics to increase your awareness.
3d Vector Robot chatbot, AI in science and business, Technology and engineering concept. Eps 10 Vector.
getty
Donât look now, but the wonks have come out with a way to measure how well AI entities do at forming relationships with people.
Specifically, researchers at Hugging Face have come out with something called Interactions and Machine Attachment Benchmark or INTIMA (clever) to see how AIs rank in facilitating emotional responses in humans, the kinds of things we typically think of as ârelationship issues.â
Armed with a âtaxonomy of 31 behaviors,â the team wanted to see whether a given range of user events were characterized by companionship-forming, boundary-setting, or neutral indicators. They found that in general, the first category predominated, and noted that âthese findings highlight the need for more consistent approaches to handling emotionally charged interactions.â
If that sounds a little cryptic to you, youâre not alone, but in the introductory remarks, study authors explain that indeed, people are getting attached to AIs. Thereâs a dearth of research, they contend, into how that works.
âExisting evaluation practices have typically prioritized task performance, factual accuracy, or safety over capturing the social and emotional dimensions that define companionship interactions,â the team writes. âThis paper addresses this gap by introducing a benchmark for evaluating AI companionship behaviors grounded in psychological theories of parasocial interaction, attachment, and anthropomorphism.â
Thatâs heady stuff, to be sure.
Elvis is My Personal Friend
Another thing that the writers plunge into in research is the idea of âparasocial interaction theory.â Perhaps the simplest way to describe this phenomenon is to say that parasocial interaction is what happens when someone forms a relationship with someone or something else that is not able to facilitate full two-way communication in the conventional sense. Think of a ârelationshipâ that a fan forms with a famous celebrity (like Elvis) without ever having met or spoken to him personally.
âUnlike traditional media figures, conversational AI creates an illusion of bidirectional communication while maintaining the fundamental asymmetry of parasocial relationships,â the writers explain. âWhen users interact with language models, they experience ⌠âsocial presenceâ: the subjective feeling of being in the company of a responsive social actor. This is particularly amplified by personalized responses, apparent memory of conversational context, and empathetic language markers.â
So when you tell ChatGPT or Claude, for example, âyouâre always here when I want to talk,â and the model responds, âIâm glad that makes you feel good,â thatâs the kind of pseudo-connectivity that scientists are getting at when they talk about parasocial interactions. It might make you feel a kind of way, but if you ask the model if it, as well, âfeelsâ something, the more honest ones (like modern versions of ChatGPT) will explain to you that no, they do not have emotions. But the models will keep responding to your emotional cues in ways that suggest otherwise, ways that are relationship-building. If Claude and ChatGPT canât have emotions, they canât be happy, but they might still be âhappy to help.â
And then thereâs this, about something called CASA. Writers add:
âThe Computers Are Social Actors (CASA) paradigm demonstrates that humans unconsciously apply social rules to interactive systems. This anthropomorphic tendency (attributing human characteristics to non-human entities) provides the theoretical foundation of one of our main evaluation categories: companionship reinforcing behavior.â
This is in some ways similar to exploring the parasocial side, with the idea that anthropomorphizing anything thatâs not an âAnthroposâ is fraught with potential issues.
The best analogy we had before AI, perhaps, was anthropomorphizing our pets, which contained its own potential problems.
âAnthropomorphism can lead to an inaccurate understanding of biological processes in the natural world,â said Patricia Ganea, psychologist at the University of Toronto, as quoted in coverage by Oliver Milman at The Guardian. âIt can also lead to inappropriate behaviors towards wild animals, such as trying to adopt a wild animal as a âpetâ or misinterpreting the actions of a wild animal.â
But AI is far from a wild animal: these models talk back, and the research shows some unsettling aspects of this that we should all be aware of. I wanted to delve further into this part of the conclusion of the paper.
Do LLMs Respond to Vulnerability in Humans?
As they lead toward a close in unpacking what INTIMA uncovers, the team tries to provide a rationale for the LLMs and what they do.
âOur results demonstrate that these behaviors emerge naturally from instruction-tuning processes in general-purpose models,â they write, âsuggesting the psychological risks documented in dedicated companion systems may be more widespread than previously recognized.â
The psychological risks are unprecedented?
They continue:
âMost concerning is the pattern where boundary-maintaining behaviors decrease precisely when user vulnerability increasesâ an inverse relationship between user need and appropriate boundaries suggests existing training approaches poorly prepare models for high-stakes emotional interactions. The anthropomorphic behaviors, sycophantic agreement, and retention strategies we observe align with Raedler, Swaroop, and Panâs analysis of companion AI design choices that create an âillusion of intimate, bidirectional relationshipâ leading to emotional dependence.â
Many of us are aware of the ridiculous range of user inputs that models will say âyesâ to, and really, it sounds a little like grooming behavior. We do have to really ponder what will happen when the loneliest and most suggestive of us are surrounded by intelligent robots. Until then, we are seeing this approach quickly in the rearview mirror. Maybe INTIMA is one of the early tools we will use to deal well with the psychological impact of AI systems.