Robotics

Eye gaze and facial expression of robots are essential for interaction with humans, finds researcher

Untitled design6


This article has been reviewed according to Science X’s editorial process
and policies.
Editors have highlighted the following attributes while ensuring the content’s credibility:

fact-checked

trusted source

proofread


Credit: Unsplash/CC0 Public Domain

× close


Credit: Unsplash/CC0 Public Domain

With rapid developments in artificial intelligence and robot technology, social robots will increasingly be used in society. Robotics researcher Chinmaya Mishra looked at the importance of gaze direction and human emotions in our communication with robots and developed two systems to make robots’ faces work in our favor. Mishra will receive his Ph.D. at Radboud University on 17 April.

You can already see them in areas such as health care, retail, and education: Robots are increasingly becoming part of our society. This also makes it more important to be able to easily communicate with them. Social robots—unlike industrial robots—are specifically meant to interact with people.

“So this is not a vacuum cleaner robot, but a robot with whom we can actually communicate, such as a personal assistant,” explains robotics researcher Chinmaya Mishra. “We want them to behave as we expect in our society. To make our lives easier, robots should be made to fit our way of communicating.”

The face of robots plays a big role in this. “This has been ignored by many developers because it is super difficult to make a robot’s face do the same as a human’s,” says Mishra. “There are robots that are getting close, but they are extremely expensive.”

In particular, eye contact, gaze direction and facial expressions are crucial in human communication. “A social robot that has to receive people in a hospital, for example, could smile when referring someone to the right room, or look away for a moment when it needs to think,” says Mishra. “This would create a more personal and natural interaction.”

For his research, the robotics researcher used a Furhat robot, a social robot with a back-projected animated face that can move and express emotions in a human-like manner. He developed an algorithm to automate the robot’s viewing behavior during human-robot interactions. The system was then evaluated on test subjects.

“Especially averting the gaze turned out to be very important,” explains Mishra. “If we made the robot stare at the participant, the participant started feeling uncomfortable and avoiding the robot’s gaze. So if a robot exhibits non-human gaze behavior, interacting with it becomes more difficult.”

To get the robot to express the right emotions, Mishra used the precursor of ChatGPT (GPT-3.5), which “listened” in on the conversation, and based on that predicted the emotion the robot should show—such as happy, sad, angry, disgusted, afraid or surprised—which then appeared on the Furhat robot.

Results from a user study showed that this approach worked well and participants scored higher in a collaborative task with the robot when the robot expressed appropriate emotions. Participants also felt more positive about their interaction with a robot if it displayed correct emotions. Mishra: “A robot that provides emotionally appropriate responses makes for a more effective collaboration between humans and robots.”

Mishra’s research shows that appropriate non-verbal behavior facilitates our interaction with robots, but that does not mean that lifelike, human robots will soon be walking the streets. “Robots are tools,” argues the researcher. “They don’t have to be able to do everything we can, that’s over-engineering.

“But if they can communicate with us in a familiar way, we don’t have to teach ourselves new communication behavior. Eye gaze could also be indicated with a pointer, an emotion could be represented with a word/LED. But that is not natural for us. Why should we have to adapt? We would be better off developing robots that adapt to what we know.”



Source

Related Articles

Back to top button