Evaluation of Human-AI Teams for Learned and Rule-Based Agents in Hanabi

Ho Chit Siu · Jaime Peña · Edenna Chen · Yutai Zhou · Victor Lopez · Kyle Palko · Kimberlee Chang · Ross Allen

Keywords: [ Reinforcement Learning and Planning ] [ Deep Learning ] [ Interpretability ]

[ Abstract ]
Thu 9 Dec 8:30 a.m. PST — 10 a.m. PST


Deep reinforcement learning has generated superhuman AI in competitive games such as Go and StarCraft. Can similar learning techniques create a superior AI teammate for human-machine collaborative games? Will humans prefer AI teammates that improve objective team performance or those that improve subjective metrics of trust? In this study, we perform a single-blind evaluation of teams of humans and AI agents in the cooperative card game Hanabi, with both rule-based and learning-based agents. In addition to the game score, used as an objective metric of the human-AI team performance, we also quantify subjective measures of the human's perceived performance, teamwork, interpretability, trust, and overall preference of AI teammate. We find that humans have a clear preference toward a rule-based AI teammate (SmartBot) over a state-of-the-art learning-based AI teammate (Other-Play) across nearly all subjective metrics, and generally view the learning-based agent negatively, despite no statistical difference in the game score. This result has implications for future AI design and reinforcement learning benchmarking, highlighting the need to incorporate subjective metrics of human-AI teaming rather than a singular focus on objective task performance.

Chat is not available.