Abstract

Deep reinforcement learning (deep RL) achieved big successes with the advantage of deep learning techniques, while it also introduces the disadvantage of the model interpretability. Bad interpretability is a great obstacle for deep RL to be applied in real situations or human-machine interaction situations. Borrowed from the deep learning field, the techniques of saliency maps recently become popular to improve the interpretability of deep RL. However, the saliency maps still cannot provide specific and clear enough model interpretations for the behavior of deep RL agents. In this paper, we propose to use hierarchical conceptual embedding techniques to introduce prior-knowledge in the deep neural network (DNN) based models of deep RL agents and then generate the saliency maps for all the embedded factors. As a result, we can track and discover the important factors that influence the decisions of deep RL agents.

Details

Title
Interpretable Saliency Map for Deep Reinforcement Learning
Author
Zheng, Hong 1 ; Dai, Yinglong 2 ; Yu, Fumin 3 ; Hu, Yuezhen 3 

 School of Physics and Electronics, Hunan Normal University, Changsha, 410081, China 
 College of Liberal Arts and Sciences, National University of Defense Technology, Changsha, 410073, China; Hunan Provincial Key Laboratory of Intelligent Computing and Language Information Processing, Hunan Normal University, Changsha, 410081, China 
 Hunan Provincial Key Laboratory of Intelligent Computing and Language Information Processing, Hunan Normal University, Changsha, 410081, China 
Publication year
2021
Publication date
Jan 2021
Publisher
IOP Publishing
ISSN
17426588
e-ISSN
17426596
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
2513035913
Copyright
© 2021. This work is published under http://creativecommons.org/licenses/by/3.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.