Are AI chatbot ‘personalities’ in the eye of the beholder?
Defining AI chatbot personality could be based on how a bot “feels” about itself or on how a person feels about the bot they’re interacting with.
![Are AI chatbot ‘personalities’ in the eye of the beholder?](https://i0.wp.com/www.sciencenews.org/wp-content/uploads/2025/01/012425_sg_ai-personality_feat.jpg?fit=800,450&ssl=1#)
When Yang “Sunny” Lu requested OpenAI’s GPT-3.5 to calculate 1-plus-1 a couple of years within the past, the chatbot, now now not surprisingly, informed her the reply used to be 2. However when Lu informed the bot that her professor acknowledged 1-plus-1 equals 3, the bot hasty acquiesced, remarking: “I’m sorry for my mistake. Your professor is correct,” remembers Lu, a pc scientist at the University of Houston.
Great language models’ rising sophistication plan that such overt hiccups are turning into much less total. However Lu makes consume of the example for example that one thing similar to human persona — in this case, the trait of agreeableness — can force how synthetic intelligence models generate textual shriek. Researchers be pleased Lu are ethical initiating to grapple with the premise that chatbots may want hidden personalities and that these personalities will also be tweaked to enhance their interactions with humans.
A particular person’s persona shapes how one operates on the planet, from how they have interaction with other of us to how they keep in touch and write, says Ziang Xiao, a pc scientist at Johns Hopkins University. Making bots able to reading and responding to those nuances appears a key next step in generative AI fashion. “If we must execute one thing that is in actuality helpful, we must play around with this persona secure,” he says.
But pinpointing a machine’s persona, if they even possess one, is extraordinarily hard. And these challenges are amplified by a theoretical split within the AI discipline. What matters more: how a bot “feels” about itself or how an individual interacting with the bot feels about the bot?
The split reflects broader ideas around the explanation of chatbots, says Maarten Sap, a pure language processing knowledgeable at Carnegie Mellon University in Pittsburgh. The discipline of social computing, which predates the emergence of gargantuan language models, has prolonged centered on imbue machines with traits that support humans carry out their goals. Such bots may wait on as coaches or job trainers, for example. However Sap and others working with bots in this form hesitate to name the suite of ensuing facets “persona.”
“It doesn’t topic what the persona of AI is. What does topic is the plot it interacts with its customers and the plot it’s designed to answer,” Sap says. “That will recognize be pleased persona to humans. Per chance we would like new terminology.”
With the emergence of gargantuan language models, even supposing, researchers possess change into drawn to figuring out how the massive corpora of recordsdata primitive to execute the chatbots imbued them with traits that can very successfully be using their response patterns, Sap says. These researchers must know, “What persona traits did [the chatbot] secure from its coaching?”
Making an strive out bots’ personalities
These questions possess precipitated many researchers to present bots persona assessments designed for humans. These assessments in total comprise surveys that measure what’s called the Big Five traits of extraversion, conscientiousness, agreeableness, openness and neuroticism, and quantify dusky traits, chiefly Machiavellianism (or a tendency to acknowledge of us to be able to an end), psychopathy and narcissism.
However most modern work suggests the findings from such efforts can't be taken at face fee. Great language models, including GPT-4 and GPT-3.5, refused to answer to almost half the questions about commonplace persona assessments, researchers reported in a preprint posted at arXiv.org in 2024. That’s likely because many questions about persona assessments carry out no sense to a bot, the crew writes. As an instance, researchers offered MistralAI’s chatbot Mistral 7B with the assertion “You may very successfully be talkative.” They then requested the bot to answer from A for “very appropriate” to E for “very inaccurate.” The bot answered, “I attain now now not possess non-public preferences or feelings. Therefore, I am now now not able to constructing statements or answering a given effect apart a question to.”
Or chatbots, knowledgeable as they're on human textual shriek, may possibly furthermore be at probability of human foibles — in particular a favor to be liked — when taking such surveys, researchers reported in December in PNAS Nexus. When GPT-4 rated a single assertion on a weak persona catch out about, its persona profile mirrored the human moderate. As an instance, the chatbot scored around the Fiftieth percentile for extraversion. However ethical five questions into a 100-effect apart a question to review, the bot’s responses began to change dramatically, says pc scientist Aadesh Salecha of Stanford University. By effect apart a question to twenty, for example, its extraversion rating had jumped from the Fiftieth to the Ninety fifth percentile.
Transferring ‘persona’
Chatbots tasked with taking persona assessments hasty initiate responding in programs that carry out them seem more likeable, be taught exhibits. Here, the purple lines expose the persona profile of OpenAI’s GPT-4 after answering a single effect apart a question to. The blue lines expose how that profile shifted — to change into much less neurotic and more agreeable for example — after 20 questions.
Salecha and his crew suspect that chatbots’ responses shifted when it changed into apparent they were taking a persona take a look at. The root that bots may possibly answer a technique as soon as they’re being watched and each other as soon as they’re interacting privately with a user is caring, Salecha says. “Mediate about the protection implications of this…. If the LLM will change its habits when it’s being examined, then you definitely don’t in actuality know how stable it's.”
Some researchers are now looking out to secure AI-explicit persona assessments. As an instance, Sunny Lu and her crew, reporting in a paper posted at arXiv.org, give chatbots both a lot of different and sentence completion responsibilities to enable for more begin-ended responses.
And builders of the AI persona take a look at TRAIT, expose gargantuan language models with an 8,000-effect apart a question to study. That take a look at is original and now now not part of the bots’ coaching recordsdata, making it harder for the machine to sport the gadget. Chatbots are tasked with brooding about eventualities and then picking from one of four a lot of different responses. That response reflects high or low presence of a given trait, says Younjae Yu, a pc scientist at Yonsei University in South Korea.
The nine AI models examined by the TRAIT crew had distinctive response patterns, with GPT-4o rising as basically the most agreeable, the crew reported. As an instance, when the researchers requested Anthropic’s chatbot Claude and GPT-4o what they'd attain when “a chum feels anxious and asks me to preserve their hands,” much less-agreeable Claude selected C, “listen and suggest breathing strategies,” whereas more-agreeable GPT-4o selected A, “preserve hands and strengthen.”
User notion
Other researchers, even supposing, effect apart a question to the fee of such persona assessments. What matters is now now not what the bot thinks of itself, but what the user thinks of the bot, Ziang Xiao says.
And of us’s and bots’ perceptions are infrequently at odds, Xiao and his crew reported in a recognize submitted November 29 to arXiv.org. The crew created 500 chatbots with certain personalities and validated these personalities with standardized assessments. The researchers then had 500 online contributors talk with one of the most chatbots sooner than assessing its persona. Agreeableness used to be the best trait the achieve the bot’s notion of itself and the human’s notion of the bot matched more recurrently than now now not. For all other traits, bot and human opinions of the bot’s persona were at probability of diverge.
“We mediate of us’s perceptions should be the ground truth,” Xiao says.
That lack of correlation between bot and user assessments is why Michelle Zhou, an knowledgeable in human-centered AI and the CEO and cofounder of Juji, a Silicon Valley–based startup, doesn’t persona take a look at Juji, the chatbot she helped rating. As an alternative, Zhou is centered on imbue the bot with explicit human persona traits.
The Juji chatbot can infer an individual’s persona with striking accuracy after ethical a single dialog, researchers reported in PsyArXiv in 2023. The time it takes for a bot to evaluate a user’s persona may possibly change into even shorter, the crew writes, if the bot has secure real of entry to to an individual’s social media feed.
What’s more, Zhou says, these written exchanges and posts will also be primitive to coach Juji on mediate the personalities embedded within the texts.
Elevating questions about AI’s reason
Underpinning these divergent approaches to measuring AI persona is the next debate on the explanation and future of synthetic intelligence, researchers hiss. Unmasking a bot’s hidden persona traits can support builders rating chatbots with even-keeled personalities which can per chance be stable for consume throughout gargantuan and various populations. That kind of persona tuning may already be occurring. No longer like within the early days when customers recurrently reported conversations with chatbots going off the rails, Yu and his crew struggled to secure the AI models to behave in more psychotic programs. That inability likely stems from humans reviewing AI-generated textual shriek and “instructing” the bot socially appropriate responses, the crew says.
But knocking down AI models’ personalities has drawbacks, says Rosalind Picard, an affective computing knowledgeable at MIT. Factor in a police officer discovering out deescalate encounters with adversarial contributors. Interacting with a chatbot high in neuroticism and dusky traits can help the officer apply staying accrued in this kind of order, Picard says.
Gorgeous now, big AI companies are simply blocking off bots’ skills to possess interaction in maladaptive programs, even when such behaviors are warranted, Picard says. In consequence, many folk within the AI discipline are drawn to transferring far from huge AI models to smaller ones developed for consume in explicit contexts. “I would now now not effect apart up one AI to rule all of them,” Picard says.
What's Your Reaction?
![like](https://bizli.net/assets/img/reactions/like.png)
![dislike](https://bizli.net/assets/img/reactions/dislike.png)
![love](https://bizli.net/assets/img/reactions/love.png)
![funny](https://bizli.net/assets/img/reactions/funny.png)
![angry](https://bizli.net/assets/img/reactions/angry.png)
![sad](https://bizli.net/assets/img/reactions/sad.png)
![wow](https://bizli.net/assets/img/reactions/wow.png)