New! Sign up for our free email newsletter.
Science News
from research organizations

Trust issues: Users more gullible when they customize their technology

Date:
January 11, 2017
Source:
Penn State
Summary:
Cheery robots may give people the creeps and serious robots may actually ease anxiety depending on how users perceive the robot's role in their lives, according to an international team of researchers.
Share:
FULL STORY

Cheery robots may give people the creeps and serious robots may actually ease anxiety depending on how users perceive the robot''s role in their lives, according to an international team of researchers.

In a study of robot use in a retirement home, senior citizens were more likely to consider an assistant robot as socially attractive and intelligent when they thought its demeanor was cheery, according to S. Shyam Sundar, Distinguished Professor of Communications and co-director of the Media Effects Research Laboratory, Penn State. However, the seniors were more willing to accept robots labeled as companions if the machine had a more serious demeanor.

"We were actually surprised to find out that they wanted companion robots to be serious and assistant robots to be playful," said Sundar. "But it''s pretty clear from our data that a serious demeanor adds credibility to a companion robot, whereas a playful demeanor softens the tension when interacting with an assistant robot."

Assistant robots are designed to help people with everyday tasks, such as dispense dispensing medications, whereas companion robots are designed to support people emotionally, serving as friends or pets, according to the researchers, who report their findings in the current issue of the International Journal of Human-Computer Studies.

Knowing how people might react to a robot''s demeanor and role could help designers make robots that people are more likely to accept and use, according to Sundar.

"The variations of role and demeanor are shown to indirectly affect robot use intentions, and are therefore important to understand, as more initiatives target senior citizens as a prime user population for social robotics," said Sundar.

Both companion and assistant robots will likely find roles in hospitals and retirement homes to assist senior citizens with the physical, mental and emotional challenges of growing old, he added.

"In the future, we might see robots working in the healthcare system to provide senior citizens with physical assistance and emotional support in the comfort of their own homes," said Sundar. "Therefore, it''s becoming more important to understand how we can promote healthy communications between senior citizens and robots."

The researchers recruited 51 senior citizens at a central Pennsylvania retirement home. The participants were assigned to one of four conditions -- -- playful assistant robot, serious assistant robot, playful companion robot and serious companion robot.

Participants were told that the robot, which was -- about 4 feet tall and is equipped with a robotic arm, , web camera and screen, -- was either a companion or an assistant. They used a smartphone to tell the robot to approach. The robot would then approach and say hello to the participant. Two vocal recordings were used to manipulate the robot's demeanor of the robot. The higher- pitched voice was considered more playful. The serious voice had limited changes in inflection and tone.

After the interaction was over, the robot moved back to its original position and the participants then filled out a questionnaire on the interaction.


Story Source:

Materials provided by Penn State. Note: Content may be edited for style and length.


Journal Reference:

  1. S. Shyam Sundar, Eun Hwa Jung, T. Franklin Waddell, Ki Joon Kim. Cheery companions or serious assistants? Role and demeanor congruity as predictors of robot attraction and use intentions among senior citizens. International Journal of Human-Computer Studies, 2017; 97: 88 DOI: 10.1016/j.ijhcs.2016.08.006

Cite This Page:

Penn State. "Trust issues: Users more gullible when they customize their technology." ScienceDaily. ScienceDaily, 11 January 2017. <www.sciencedaily.com/releases/2017/01/170111151831.htm>.
Penn State. (2017, January 11). Trust issues: Users more gullible when they customize their technology. ScienceDaily. Retrieved November 22, 2024 from www.sciencedaily.com/releases/2017/01/170111151831.htm
Penn State. "Trust issues: Users more gullible when they customize their technology." ScienceDaily. www.sciencedaily.com/releases/2017/01/170111151831.htm (accessed November 22, 2024).

Explore More

from ScienceDaily

RELATED STORIES