TY - GEN
T1 - Leveraging domain knowledge to learn normative behavior
T2 - 2011 Adaptive and Learning Agents Workshop, ALA 2011, Held at Autonomous Agents and Multi-Agent Systems Conference, AAMAS 2011
AU - Hosseini, Hadi
AU - Ulieru, Mihaela
N1 - Copyright:
Copyright 2012 Elsevier B.V., All rights reserved.
PY - 2012
Y1 - 2012
N2 - This paper addresses the problem of norm adaptation using Bayesian reinforcement learning. We are concerned with the effectiveness of adding prior domain knowledge when facing environments with different settings as well as with the speed of adapting to a new environment. Individuals develop their normative framework via interaction with their surrounding environment (including other individuals). An agent acquires the domain-dependent knowledge in a certain environment and later reuses them in different settings. This work is novel in that it represents normative behaviors as probabilities over belief sets. We propose a two-level learning framework to learn the values of normative actions and set them as prior knowledge, when agents are confident about them, to feed them back to their belief sets. Developing a prior belief set about a certain domain can improve an agent's learning process to adjust its norms to the new environment's dynamics. Our evaluation shows that a normative agent, having been trained in an initial environment, is able to adjust its beliefs about the dynamics and behavioral norms in a new environment. Therefore, it converges to the optimal policy more quickly, especially in the early stages of learning.
AB - This paper addresses the problem of norm adaptation using Bayesian reinforcement learning. We are concerned with the effectiveness of adding prior domain knowledge when facing environments with different settings as well as with the speed of adapting to a new environment. Individuals develop their normative framework via interaction with their surrounding environment (including other individuals). An agent acquires the domain-dependent knowledge in a certain environment and later reuses them in different settings. This work is novel in that it represents normative behaviors as probabilities over belief sets. We propose a two-level learning framework to learn the values of normative actions and set them as prior knowledge, when agents are confident about them, to feed them back to their belief sets. Developing a prior belief set about a certain domain can improve an agent's learning process to adjust its norms to the new environment's dynamics. Our evaluation shows that a normative agent, having been trained in an initial environment, is able to adjust its beliefs about the dynamics and behavioral norms in a new environment. Therefore, it converges to the optimal policy more quickly, especially in the early stages of learning.
UR - http://www.scopus.com/inward/record.url?scp=84858017646&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84858017646&partnerID=8YFLogxK
U2 - 10.1007/978-3-642-28499-1_5
DO - 10.1007/978-3-642-28499-1_5
M3 - Conference contribution
AN - SCOPUS:84858017646
SN - 9783642284984
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 70
EP - 84
BT - Adaptive and Learning Agents - International Workshop, ALA 2011, Held at AAMAS 2011, Revised Selected Papers
Y2 - 2 May 2011 through 2 May 2011
ER -