TY - CHAP
T1 - Pragmatic-Pedagogic Value Alignment
AU - Fisac, Jaime F.
AU - Gates, Monica A.
AU - Hamrick, Jessica B.
AU - Liu, Chang
AU - Hadfield-Menell, Dylan
AU - Palaniappan, Malayandi
AU - Malik, Dhruv
AU - Sastry, S. Shankar
AU - Griffiths, Thomas L.
AU - Dragan, Anca D.
N1 - Publisher Copyright:
© 2020, Springer Nature Switzerland AG.
PY - 2020
Y1 - 2020
N2 - As intelligent systems gain autonomy and capability, it becomes vital to ensure that their objectives match those of their human users; this is known as the value-alignment problem. In robotics, value alignment is key to the design of collaborative robots that can integrate into human workflows, successfully inferring and adapting to their users’ objectives as they go. We argue that a meaningful solution to value alignment must combine multi-agent decision theory with rich mathematical models of human cognition, enabling robots to tap into people’s natural collaborative capabilities. We present a solution to the cooperative inverse reinforcement learning (CIRL) dynamic game based on well-established cognitive models of decision making and theory of mind. The solution captures a key reciprocity relation: the human will not plan her actions in isolation, but rather reason pedagogically about how the robot might learn from them; the robot, in turn, can anticipate this and interpret the human’s actions pragmatically. To our knowledge, this work constitutes the first formal analysis of value alignment grounded in empirically validated cognitive models.
AB - As intelligent systems gain autonomy and capability, it becomes vital to ensure that their objectives match those of their human users; this is known as the value-alignment problem. In robotics, value alignment is key to the design of collaborative robots that can integrate into human workflows, successfully inferring and adapting to their users’ objectives as they go. We argue that a meaningful solution to value alignment must combine multi-agent decision theory with rich mathematical models of human cognition, enabling robots to tap into people’s natural collaborative capabilities. We present a solution to the cooperative inverse reinforcement learning (CIRL) dynamic game based on well-established cognitive models of decision making and theory of mind. The solution captures a key reciprocity relation: the human will not plan her actions in isolation, but rather reason pedagogically about how the robot might learn from them; the robot, in turn, can anticipate this and interpret the human’s actions pragmatically. To our knowledge, this work constitutes the first formal analysis of value alignment grounded in empirically validated cognitive models.
KW - Dynamic game theory
KW - Human-robot interaction
KW - Value alignment
UR - http://www.scopus.com/inward/record.url?scp=85107065248&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85107065248&partnerID=8YFLogxK
U2 - 10.1007/978-3-030-28619-4_7
DO - 10.1007/978-3-030-28619-4_7
M3 - Chapter
AN - SCOPUS:85107065248
T3 - Springer Proceedings in Advanced Robotics
SP - 49
EP - 57
BT - Springer Proceedings in Advanced Robotics
PB - Springer Science and Business Media B.V.
ER -