Utilizing AI tools for composing messages to friends might not be the best choice, especially if the friend learns about the AI’s involvement, recent research indicates. The study revealed that participants felt a fictional friend using AI for message crafting appeared less sincere in their efforts compared to one who crafted their message manually.
That perception may be understandable, but the implications extend beyond just the content of the message, expressed Bingjie Liu, the study’s principal author and an assistant professor of communication at The Ohio State University.
“After they get an AI-assisted message, people feel less satisfied with their relationship with their friend and feel more uncertain about where they stand,” Liu said.
But to be fair to AI, it wasn’t just the use of technology that turned people off. The study also found negative effects when people learned their friend got help from another person to write a message.
“People want their partners or friends to put forth the effort to come up with their own message without help – from AI or other people,” Liu said.
The study was published online recently in the Journal of Social and Personal Relationships.
As AI chatbots like ChatGPT become increasingly popular, issues about how to use them will become more relevant and complex, Liu said.
The study involved 208 adults who participated online. Participants were told that they had been good friends with someone named Taylor for years. They were given one of three scenarios: They were experiencing burnout and needed support, they were having a conflict with a colleague and needed advice, or their birthday was coming up.
Participants were then told to write a short message to Taylor describing their current situation in a textbox on their computer screen.
All participants were told Taylor sent them a reply. In the scenarios, Taylor wrote an initial draft. Some participants were told Taylor had an AI system to help revise the message to achieve the proper tone, others were told a member of a writing community helped make revisions, and a third group was told Taylor made all edits to the message.
In every case, people in the study were told the same thing about Taylor’s reply, including that it was “thoughtful.” Still, participants in the study had different views about the message they had supposedly received from Taylor. Those who received a reply helped by AI rated what Taylor did as less appropriate and more improper than did those who received the reply that was written only by Taylor.
AI replies also led participants to express less satisfaction with their relationship, such as rating Taylor lower on meeting “my needs as a close friend.”
In addition, people in the study were more uncertain about their relationship with Taylor if they received the AI-aided response, being less certain about the statement “Taylor likes me as a close friend.”
One possible reason that people may not like the AI-aided response could be that people think using technology is inappropriate and inferior to humans in crafting personal messages like these.
But results showed that people responded just as negatively to responses in which Taylor had another human – a member of an online writing community – help with the message.
“What we found is that people don’t think a friend should use any third party – AI or another human – to help maintain their relationship,” Liu said.
The reason, the study found, was that participants felt Taylor expended less effort on their relationship by relying on AI or another person to help craft a message.
The lower participants rated Taylor’s effort by using AI or another person, the less satisfied they were with their relationship and the more uncertainty they felt about the friendship.
“Effort is very important in a relationship,” Liu said. “People want to know how much you are willing to invest in your friendship and if they feel you are taking shortcuts by using AI to help, that’s not good.”
Of course, most people won’t tell a friend that they used AI to help craft a message, Liu said. But she noted that as ChatGPT and other services become more popular, people may start doing a Turing Test in their minds as they read messages from friends and others.
The phrase “Turing Test” is sometimes used to refer to people wondering if they can tell whether an action was taken by a computer or a person.
“It could be that people will secretly do this Turing Test in their mind, trying to figure out if messages have some AI component,” Liu said. “It may hurt relationships.”
The answer is to do your own work in relationships, she said.
“Don’t use technology just because it is convenient. Sincerity and authenticity still matter a lot in relationships.”
Reference: “Artificial intelligence and perceived effort in relationship maintenance: Effects on relationship satisfaction and uncertainty” by Bingjie Liu, Jin Kang and Lewen Wei, 18 July 2023, Journal of Social and Personal Relationships.
DOI: 10.1177/02654075231189899
Liu conducted the study with Jin Kang of Carleton University in Canada and Lewen Wei of the University of New South Wales in Australia.
>>> Read full article>>>
Copyright for syndicated content belongs to the linked Source : SciTechDaily – https://scitechdaily.com/scientists-reveal-why-using-chatgpt-to-message-your-friends-isnt-a-good-idea/