acm-header
Sign In

Communications of the ACM

ACM Careers

Researchers Promote Cooperation Between Humans, Autonomous Machines


View as: Print Mobile App Share:
emotion expressions of virtual faces

Human bias against machines is mitigated by social cues such as ethnicity, displayed in expressions of virtual faces for participants in the United States (top row) and Japan (bottom).

Credit: PLOS One

The trust between humans and autonomous machines is a top priority for U.S. Army researchers — as machines become integral to society, it is critical to understand the impact on human decision-making.

Celso M. de Melo, a computer scientist from the U.S. Army Combat Capabilities Development Command's Army Research Laboratory at its ARL West campus in Playa Vista, Calif., along with Japan's Gifu University Professor Kazunori Terada, share common research interests, in particular, on human-machine cooperation and the social effects of emotion in human-machine interaction. Their research shows people readily engage in social categorization distinguishing humans from machines, which leads to reduced cooperation with machines.

A simple cultural cue — the ethnicity of the machine's virtual face — mitigated this bias for participants from two distinct cultures (Japan and United States), the researchers say.

The researchers describe their work in "Cooperation With Autonomous Machines Through Culture and Emotion," published in the jouranl PLOS One.

"Our research shows that situational cues of affiliative intent — namely, expressions of emotion — overrode expectations of coalition alliances from social categories," de Melo says. "When machines were from a different culture, participants showed the usual bias when competitive emotion was shown — such as joy following exploitation and in contrast, participants cooperated just as much with humans as machines that expressed cooperative emotion — such as joy following cooperation."

A fundamental challenge for the adoption of AI in the Army, this research is particularly relevant for the Army's Next Generation Combat Vehicle modernization priority, as it proposes solutions to promote effective human-machine teaming, de Melo says.

"Research reveals a fundamental bias, whereby humans make decisions with machines that are less favorable than if the partners were humans," de Melo says. "It is as if machines are treated, by default, as out-group members. The research advances two solutions for overcoming this bias based on social categorization: associating positive social groups with machines and emotion; and expression of emotion by machines that communicate an affiliative intent."

The next step is to apply and test these findings in Army-relevant domains, such as warfighter interaction with autonomous ground and air vehicles, robots, and virtual partners, he says.


 

No entries found

Sign In for Full Access
» Forgot Password? » Create an ACM Web Account