UNIVERSITY PARK, Pa. — A Penn State-led research team received a $2 million National Science Foundation (NSF) grant to study how machines perceive and process human body language. The research could stimulate advancements in health care applications, including the design of caregiving robots.
The grant is part of the NSF’s Computer and Information Science and Engineering Community Research Infrastructure program. Penn State researchers will collaborate with the University of Illinois at Chicago (UIC) and the Robotics, Automation and Dance (RAD) Lab in Philadelphia.
“Extensive research has been conducted on how the human face conveys emotion, but more research is needed to understand how the body communicates feelings,” said James Wang, distinguished professor of information sciences and technology at Penn State, who is leading the project. “Our work training artificial intelligence technologies to identify patterns in the bodily movements that accompany emotions may provide essential insights for shaping the future of human-machine interaction.”
Wang will work alongside Penn State collaborators Reginald Adams Jr., professor of psychology; Michelle Newman, professor of psychology; and Jia Li, professor of statistics.
The multidisciplinary team will begin the project by exploring the wealth of information about human movement that already exists in films and internet videos. Analyzing tens of thousands of online clips culled from these videos will enable the researchers to compile a large dataset that could lead to a representative evaluation of how actors portray emotions in various contexts. For example, bowing the head with a sinking sense along the spine may be correlated with an expression of sadness. The Penn State team pioneered the data collection process in 2019.
“The subtlety and complexity of human expressions, coupled with the challenges of privacy issues, make it next to impossible to conduct this research ‘in the wild,’ so to speak,” Wang said. “We acknowledge that films use actors and that social media presenters are often self-filtered, but we expect realistic patterns to emerge from the sheer volume of online content we are studying.”
The research team will show the collected video clips to a diverse group of participants and ask them to identify emotions based on the bodily movements they observe. People perceive interactions differently, according to Wang, and culture, age, gender and other variables may result in different interpretations of the same emotional display.