HMClab

Human-Machine Communication Lab

Learn More

About Us


Technology

As information communication technology continues to develop, we are increasingly interacting with intelligent machines such as social robots, voice agents, chatbots, and autonomous vehicles. We directly interact with the technologies surrounding us, and digital entities have been and continue to stand in for humans in everyday communication contexts.

Communication

We study communication between humans and machines designed to enact the role of communicator. In our human-machine communication (HMC) research, technology is conceptualized as more than a channel or medium: it enters into the role of a communicator, and the messages flow between and among humans and machines.

Collaboration

Communication plays a vital role in human-machine collaboration, by enabling both parties to leverage each other’s strengths, coordinate their actions, and help each other to solve problems. Our endeavor is to develop theories that help us understand the fundamental effects of communication and collaboration between humans and machines.

Meet the Team


...

Austin Lee, Ph.D.

Associate Professor
Chapman University

...

Jake Liang, Ph.D.

Assistant Professor
Chapman University

...

Kerk Kee, Ph.D.

Associate Professor
Texas Tech University

...

Nao Evolution V6

Humanoid Robot
Softbank Robotics

...

Double

Telepresence Robot
Double Robotics

Research Highlights


Robots can Persuade People

A robot can effectively get people to do as the robot asks by capitalizing on the norm of reciprocity and following with a verbal request. We found that 60% of participants who were helped by a robot in a trivia game just for five minutes returned the robot’s favor by completing boring tasks for 15 minutes, no matter they liked the robot or not. This compliance rate was significantly higher than that of those who were not helped by the robot (33%).

Why Do People Comply?

Why do people treat robots and computers as if they were humans? Probably because people operate on a kind of autopilot when they interact with technology. In our study, a computer agent requested participants to do boring tasks. Mindless participants compiled when any kind of reason was given with the magic word “because.” Attentive participants complied only when a proper reason was given.

Fear of Robots

Using a nationally representative sample, we found that 26% of US population report a heightened level of fear toward autonomous robots and artificial intelligence. This fear was connected to sex, age, education, and household income, as well as media exposure to science fiction. It is also related to other types of fear, such as loneliness, becoming unemployed, and drone use.

Enhancing Human-Robot Trust

We examined the effects of user-generated content on human-robot trust as well as interaction outcomes. We found that reading online reviews about robot partners before interaction augmented the effects of human-robot trust on interaction outcomes. Specifically, it elicited positive mood and more favorable evaluations of the robot. This finding suggests that messages can be strategically deployed to enhance trust and interaction outcomes.

Press Coverage


From factories, offices and medical centers to our homes, cars and even our local coffee shops, robots and artificial intelligence are playing an increasingly greater role in how we work, live and play...

Cincinnati Public Radio
October 3, 2017

As social robots permeate wide segments of our daily lives, they may become capable of influencing our attitudes, beliefs, and behaviors. In other words, robots may be able to persuade us, and this is already happening...

It’s no secret that the College of Informatics has the most technologically advanced building on campus — Griffin Hall. But Griffin Hall is also home to Pineapple. Dr. Lee and Dr. Liang are using Pineapple to conduct studies on how humans interact socially when a robot is present...

NKU Inside
October, 2015

Between the studies on digital media production and computer sciences, Griffin Hall is home to some impressive technology. You might start seeing a new, digital face wandering the halls of the College of Informatics. That digital face is known as Pineapple, and it will be used for the next three years for research into social robotics...

The Northerner
October 19, 2015

Awards


Awards

Top Paper Award, National Communication Association, Communication and the Future Division (Lee & Liang, 2017)

Valerie Scudder Award: Jake Liang, Chapman University (2017)

Excellence in Research, Scholarship, and Creative Activity: Austin Lee, Northern Kentucky University (2017)

Faculty Research Excellence Award: Jake Liang, Chapman University (2016)

Top Paper Awards, International Communication Association, Instructional and Developmental Communication Division (Liang, 2015)

Grants

Pedagogical Innovations Grant (Social Robotics): Austin Lee, Chapman University (2019)

Faculty Summer Fellowship (Robot Credibiilty): Austin Lee, Northern Kentucky University (2015)

Faculty Project Grant (Social Robotics): Austin Lee, Northern Kentucky University (2015)

Research Grant (Participatory Websites): Jake Liang, Chapman University (2015)

Seed Grant (Persuasive Robotics): Austin Lee, Northern Kentucky University (2015)

Community Outreach


Publications


Journal Articles

Lee, S. A., & Liang, Y. (2019). A communication model of human-robot trust development for inclusive education. In J. Knox, Y. Wang, & M. S. Gallagher (Eds.), Speculative futures for artificial intelligence and educational inclusion (pp. 101-115). New York, NY: Springer Nature.

Lee, S. A., & Liang, Y. (2019). Robotic foot-in-the-door: Using sequential-request persuasive strategies in human-robot interaction. Computers in Human Behavior, 90, 351-356.

Lee, S. A., & Liang, Y. (2018). Theorizing verbally persuasive robots. In A. L. Guzman (Ed.), Human-machine communication: Rethinking communication, technology, and ourselves (pp. 119-143). New York, NY: Peter Lang.

Lee, S. A., & Liang, Y. (2018). Robotic foot-in-the-door: Using sequential-request persuasive strategies in human-robot interaction. Computers in Human Behavior.

Liang, Y., & Lee, S. A. (2017). Fear of autonomous robots: Evidence from national representative data with probability sampling. International Journal of Social Robotics, 9, 379-384.

Liang, Y., & Lee, S. A. (2016). Advancing the strategic messages affecting robot trust effect: The dynamic of user- and robot-generated content on human-robot trust and interaction outcomes. Cyberpsychology, Behavior, and Social Networking, 19, 538-544.

Lee, S. A., & Liang, Y. (2016). The role of reciprocity in verbally persuasive robots. Cyberpsychology, Behavior, and Social Networking. 19, 524-527.

Lee. S. A., & Liang, Y. (2015). Reciprocity in computer-human interaction: Source-based, norm-based, and affect-based explanations. Cyberpsychology, Behavior, and Social Networking, 18, 234-240.

Liang, Y., Lee, S. A., & Jang, J. (2013). Mindlessness and gaining compliance in computer-human interaction. Computers in Human Behavior, 29, 1572-1579.

Conference Presentations

Lee, S. A. (2019). Human-robot proxemics and compliance gaining. Paper presented at the 69th annual convention of the International Communication Association, Washington, DC.

Liang, Y., Lee, S. A., & Kee, K. F. (2019). The adoption of collaborative robots toward ubiquitous diffusion: A research agenda. Paper presented at the 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI 2019) Workshop, Daegu, South Korea.

Lee, S. A., Appelman, A., & Waldridge, Z. J. (2018). Robot message credibility. Paper presented at the 68th annual convention of the International Communication Association, Prague, Czech.

Lee, S. A., & Liang, Y. (2017). Theorizing verbally persuasive robots. Paper presented at the 103rd annual convention of the National Communication Association, Dallas, TX.

Lee, S. A., Liang, Y., & Thompson, A. M. (2017). Robotic foot-in-the-door: Using sequential-request persuasive strategies in human-robot interaction. Paper presented at the 67th annual convention of the International Communication Association, San Diego, CA.

Liang, Y., & Lee, S. A. (2016). Employing user-generated content to enhance human-robot interaction in a human-robot trust game. In Proceedings of the 11th ACM/IEEE International Conference on Human-Robot Interaction (HRI 2016). Christchurch, New Zealand.

Lee, S. A., Liang, Y., & Cho, S. (2016). Effects of anthropomorphism and reciprocity in persuasive computer agents. Paper presented at the at the 102nd annual convention of the National Communication Association, Philadelphia, PA.

Cho, S., Lee, S. A., & Liang, Y. (2016). Using anthropomorphic agents for persuasion. Paper presented at the 66th annual convention of the International Communication Association, Fukuoka, Japan.

Contact Us


Contact info

School of Communication
One University Drive
Orange, CA 92866

humanmachinecommunication@gmail.com

+1 714 744 7088