TY - GEN
T1 - Can't hide your disappointment
T2 - 2021 IEEE International Conference on Advanced Robotics and Its Social Impacts, ARSO 2021
AU - Surendran, Vidullan
AU - Wagner, Alan R.
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021/7/8
Y1 - 2021/7/8
N2 - Recognising intent in collaborative human robot tasks can improve team performance and human perception of the robot. Tasks that involve dynamic physical motions increase the likelihood of mistakes committed by the human. When a mistake is made, the observed outcome can differ from the intent of the human. We setup a throwing task consisting of 9 targets, and propose a method that can predict human intent in the presence of mistakes. This method uses a vision based pipeline to predict the outcome of the throw, determine if the subject's emotional reaction to the outcome indicates incongruence between the intent and outcome, and finally predict the intent of the human in the context of the throwing task. We show that the use of human pose improves outcome prediction accuracy to 28% compared to prior work that uses a two-stream architecture to achieve 22%. The method is also able to predict intent-outcome congruence accurately in 75% of the cases. Since the prediction of intent in the presence of mistakes is currently understudied, we compare against a random baseline of 11% and find that the end-to-end intent recognition pipeline achieves an accuracy of 23%.
AB - Recognising intent in collaborative human robot tasks can improve team performance and human perception of the robot. Tasks that involve dynamic physical motions increase the likelihood of mistakes committed by the human. When a mistake is made, the observed outcome can differ from the intent of the human. We setup a throwing task consisting of 9 targets, and propose a method that can predict human intent in the presence of mistakes. This method uses a vision based pipeline to predict the outcome of the throw, determine if the subject's emotional reaction to the outcome indicates incongruence between the intent and outcome, and finally predict the intent of the human in the context of the throwing task. We show that the use of human pose improves outcome prediction accuracy to 28% compared to prior work that uses a two-stream architecture to achieve 22%. The method is also able to predict intent-outcome congruence accurately in 75% of the cases. Since the prediction of intent in the presence of mistakes is currently understudied, we compare against a random baseline of 11% and find that the end-to-end intent recognition pipeline achieves an accuracy of 23%.
UR - http://www.scopus.com/inward/record.url?scp=85116473850&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85116473850&partnerID=8YFLogxK
U2 - 10.1109/ARSO51874.2021.9541546
DO - 10.1109/ARSO51874.2021.9541546
M3 - Conference contribution
AN - SCOPUS:85116473850
T3 - Proceedings of IEEE Workshop on Advanced Robotics and its Social Impacts, ARSO
SP - 21
EP - 28
BT - 2021 IEEE International Conference on Advanced Robotics and Its Social Impacts, ARSO 2021
PB - IEEE Computer Society
Y2 - 8 July 2021 through 10 July 2021
ER -