TY - GEN
T1 - Visualizing and understanding atari agents
AU - Greydanus, Sam
AU - Koul, Anurag
AU - Dodge, Jonathan
AU - Fern, Alan
N1 - Funding Information:
This material is based upon work supported by the Defense Advanced Research Projects Agency (DARPA) under Contract N66001-17-2-4030.
Publisher Copyright:
© 2018 35th International Conference on Machine Learning, ICML 2018. All rights reserved.
PY - 2018
Y1 - 2018
N2 - While deep reinforcement learning (deep RL) agents are effective at maximizing rewards, it is r often unclear what strategies they use to do so. In this paper, we take a step toward explaining deep RL agents through a case study using Atari 2600 environments. In particular, we focus on using saliency maps to understand how an agent learns and executes a policy. We introduce a method for ? generating useful saliency maps and use it to show 1) what strong agents attend to, 2) whether agents are making decisions for the right or wrong reasons, and 3) how agents evolve during learning. • We also test our method on non-expert human ∗ subjects and find that it improves their ability to reason about these agents. Overall, our results show that saliency information can provide significant insight into an RL agent's decisions and learning behavior.
AB - While deep reinforcement learning (deep RL) agents are effective at maximizing rewards, it is r often unclear what strategies they use to do so. In this paper, we take a step toward explaining deep RL agents through a case study using Atari 2600 environments. In particular, we focus on using saliency maps to understand how an agent learns and executes a policy. We introduce a method for ? generating useful saliency maps and use it to show 1) what strong agents attend to, 2) whether agents are making decisions for the right or wrong reasons, and 3) how agents evolve during learning. • We also test our method on non-expert human ∗ subjects and find that it improves their ability to reason about these agents. Overall, our results show that saliency information can provide significant insight into an RL agent's decisions and learning behavior.
UR - http://www.scopus.com/inward/record.url?scp=85057331234&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85057331234&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85057331234
T3 - 35th International Conference on Machine Learning, ICML 2018
SP - 2877
EP - 2886
BT - 35th International Conference on Machine Learning, ICML 2018
A2 - Dy, Jennifer
A2 - Krause, Andreas
PB - International Machine Learning Society (IMLS)
T2 - 35th International Conference on Machine Learning, ICML 2018
Y2 - 10 July 2018 through 15 July 2018
ER -