Abstract
This paper addresses the problem of norm adaptation using Bayesian reinforcement learning. We are concerned with the effectiveness of adding prior domain knowledge when facing environments with different settings as well as with the speed of adapting to a new environment. Individuals develop their normative framework via interaction with their surrounding environment (including other individuals). An agent acquires the domain-dependent knowledge in a certain environment and later reuses them in different settings. This work is novel in that it represents normative behaviors as probabilities over belief sets. We propose a two-level learning framework to learn the values of normative actions and set them as prior knowledge, when agents are confident about them, to feed them back to their belief sets. Developing a prior belief set about a certain domain can improve an agent’s learning process to adjust its norms to the new environment’s dynamics. Our evaluation shows that a normative agent, having been trained in an initial environment, is able to adjust its beliefs about the dynamics and behavioral norms in a new environment. Therefore, it converges to the optimal policy more quickly, especially in the early stages of learning.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Bellman, R.: Dynamic Programming. Princeton UP, Princeton (1957)
Bellman, R.: Adaptive control processes: a guided tour, 1:2. Princeton University Press (1961)
Bertsekas, D.: Dynamic programming: deterministic and stochastic models. Prentice-Hall, Inc., Upper Saddle River (1987)
Boella, G., Lesmo, L.: Deliberate normative agents. Kluwer, Norwell (2001)
Boman, M.: Norms in artificial decision making. Artificial Intelligence and Law 7(1), 17–35 (1999)
Briggs, W., Cook, D.: Flexible social laws. In: International Joint Conference on Artificial Intelligence, vol. 14, pp. 688–693. Lawrence Erlbaum Associates Ltd. (1995)
Castelfranchi, C., Dignum, F., Jonker, C., Treur, J.: Deliberative normative agents: Principles and architecture. In: Intelligent Agents VI. Agent Theories Architectures, and Languages, pp. 364–378 (2000)
Chalkiadakis, G., Boutilier, C.: Coalitional bargaining with agent type uncertainty. In: Proc. 20th IJCAI (2007)
Conte, R., Castelfranchi, C.: Cognitive and social action. Garland Science (1995)
Conte, R., Castelfranchi, C., Dignum, F.P.M.: Autonomous Norm Acceptance. In: Papadimitriou, C., Singh, M.P., Müller, J.P. (eds.) ATAL 1998. LNCS (LNAI), vol. 1555, pp. 99–112. Springer, Heidelberg (1999)
Dearden, R., Friedman, N., Andre, D.: Model based Bayesian exploration. In: Proceedings of the Fifteenth Conference on Uncertainty in Artificial Intelligence, pp. 150–159. Citeseer (1999)
Dearden, R., Friedman, N., Russell, S.: Bayesian Q-learning. In: Proceedings of the National Conference on Artificial Intelligence, pp. 761–768. John Wiley & Sons Ltd. (1998)
DeGroot, M.: Optimal statistical decisions. Wiley-IEEE (2004)
Dignum, F.: Autonomous agents with norms. Artificial Intelligence and Law 7(1), 69–79 (1999)
Dignum, F., Dignum, V.: Emergence and enforcement of social behavior. In: 8th World IMACS Congress and MODSIM 2009 International Congress on Modelling and Simulation. Modelling and Simulation Society of Australia and New Zealand and International Association for Mathematics and Computers in Simulation, pp. 2377–2383 (2009)
Epstein, J.: Learning to be thoughtless: Social norms and individual computation. Computational Economics 18(1), 9–24 (2001)
Hosseini, H.: A Reinforcement Learning Approach to Dynamic Norm Generation. Master’s thesis, University of New Brunswick (2010)
Lewis, D.: Convention: A philosophical study. Wiley-Blackwell (2002)
Martin, J., O.R.S. of America: Bayesian decision problems and Markov chains. Wiley, New York (1967)
Morris, A., Ross, W., Hosseini, H., Ulieru, M.: Modeling Culture with Complex, Multidimensional, Multiagent Systems. In: Dignum, V., Dignum, F., Ferber, J., Stratulat, T. (eds.) Integrating Cultures: Formal Models and Agent-Based Simulations. Springer Series on the Philosophy of Sociality (2011) (in print)
Mukherjee, P., Sen, S., Airiau, S.: Emergence of Norms with Biased Interactions in Heterogeneous Agent Societies. In: Proceedings of the 2007 IEEE/WIC/ACM International Conferences on Web Intelligence and Intelligent Agent Technology-Workshops, pp. 512–515. IEEE Computer Society (2007)
Sen, S., Airiau, S.: Emergence of norms through social learning. In: Proceedings of the Twentieth International Joint Conference on Artificial Intelligence, pp. 1507–1512 (2007)
Strens, M.: A Bayesian framework for reinforcement learning. In: Machine Learning-International Workshop then Conference, pp. 943–950. Citeseer (2000)
Tuomela, R.: The importance of us: A philosophical study of basic social notions. Stanford Univ. Pr. (1995)
Verhagen, H.: Norms and artificial agents. In: Sixth Meeting of the Special Interest Group on Agent-Based Social Simulation, ESPRIT Network of Excellence on Agent-Based Computing (2001)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2012 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Hosseini, H., Ulieru, M. (2012). Leveraging Domain Knowledge to Learn Normative Behavior: A Bayesian Approach. In: Vrancx, P., Knudson, M., GrzeĹ›, M. (eds) Adaptive and Learning Agents. ALA 2011. Lecture Notes in Computer Science(), vol 7113. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-28499-1_5
Download citation
DOI: https://doi.org/10.1007/978-3-642-28499-1_5
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-28498-4
Online ISBN: 978-3-642-28499-1
eBook Packages: Computer ScienceComputer Science (R0)