Skip to main content
Log in

Concept learning games

An ontological study in multi-agent systems

  • Published:
Information Systems Frontiers Aims and scope Submit manuscript

Abstract

In this paper, we intend to have a game theoretic study on the concept learning problem in a multi-agent system. Concept learning is a very essential and well-studied domain of machine learning when it is studied under the characteristics of a multi-agent system. The most important reasons are the partiality of the environment perception for any agent and also the communication holdbacks, resulting into a deep need for a collaborative protocol in favor of multi-agent transactions. Here we wish to investigate multi-agent concept learning with the help of its components, thoroughly with a game theoretic taste, esp. on the pre-learning processes. Based on two standard notations, we address the non-unanimity of concepts, classification of objects, voting and communicating protocol, and also the learning itself. In such a game of concept learning, we consider a group of agents, communicating and consulting to upgrade their ontologies based on their conceptualizations of the environment. For this purpose, we investigate the problem in two separate and standard distinctions of game theory study, cooperation and competition. Several solution concepts and innovative ideas from the multi-agent realm are used to produce an approach that contains the reasoning process of the agents in this system. Some experimentations come at the end to show the functionality of our approach. These experimentations come distinctly for both cooperative and competitive views.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

References

  • Afsharchi, M. (2007). Ontology guided collaborative concept learning in multi-agent systems. Ph.D. Dissertation, Faculty of Electrical and Computer Engineering, University of Calgary, Calgary, Alberta, Canada.

  • Afsharchi, M., Far, B. H., & Denzinger, J. (2006). Ontology-guided learning to improve communication between groups of agents. In Proc. AAMAS 2006, Hakodate (pp. 923–930).

  • Bellifemine, F. L., Caire, G., & Greenwood, D. (2007). Developing multi-agent systems with JADE. John Wiley and Sons.

  • Binmore, K. (1992). Fun and games: A text on game theory. Lexington, MA: D.C. Heath and Company.

    Google Scholar 

  • Çensoy, M., & Yolum, P. (2009). Concept learning for achieving personalized ontologies: An active learning approach. In ADMI (Vol. 5680, pp. 170–182). Springer.

  • Chalkiadakis, G., Elkind, E., & Jennings, N. R. (2009). Simple coalitional games with beliefs. In Proc. 21st international joint conference on artificial intelligence (IJCAI-09).

  • Didandeh, A., Mirbakhsh, N., & Afsharchi, M. (2010). Negotiating to learn non-unanimous concepts. In Proc. 2nd international working conference on human factors and computational models in negotiation, HuCom’10 (pp. 45–49).

  • Ephrati, E., & Rosenschein, J. S. (1993). Multi-agent planning as a dynamic search for social consensus. In Proc. 13th international joint conference on artifical intelligence (Vol. 1, pp. 423–429).

  • Finin, T., Weber, J., Wiederhold, G., Gensereth, M., Fritzzon, R., McKay, D., et al. (1993). Specification of the KQML agent-communication language. DARPA Knowledge Sharing Initiative External Interfaces Working Group.

  • Fisher, D. H. (1987). Knowledge acquisition via incremental conceptual clustering. Machine learning (Vol 2, pp. 139–172). Boston: Kluwer Academic Publishers.

  • Foundation for Intelligent Physical Agents (1999). Specification part 2-Agent Communication. http://www.fipa.org. Accessed November 2009.

  • Girgin, S., Polat, F., & Alhajj, R. (2007). Positive impact of state similarity on reinforcement learning performance. IEEE Transactions on Systems, Man, and Cybernetics-B, 37(5), 1256–1270.

    Article  Google Scholar 

  • Hendler, J. A. (2003). National Cancer Institute’s ontology of cancer, DAML, DARPA. http://www.mindswap.org/2003/CancerOntology/nciOncology.owl/. Accessed on 5 March 2010.

  • Ieong, S., & Shoham, Y. (2008). Bayesian coalitional games. In Proc. 23rd AAAI conference on artificial intelligence.

  • Illinois Semantic Integration Archive (2005). List of offered courses, computer science. http://anhai.cs.uiuc.edu/archive/. Accessed 30 January 2005.

  • Jennings, N. R., Corera, J. M., Laresgoiti, I., Mamdani, E. H., Perriollat, F., Skarek, P., et al. (1996). Using ARCHON to develop real-world DAI applications for electricity transportation management and particle accelerator control. IEEE Expert, 11(6), 60–88.

    Article  Google Scholar 

  • Kaya, M., & Alhajj, R. (2005). A novel approach to multiagent reinforcement learning: Utilizing OLAP mining in the learning process. IEEE Transactions on Systems, Man, and Cybernetics-C, 35(4), 582–590.

    Article  Google Scholar 

  • Lee, C., Jian, Z., & Huang, L. (2005). A fuzzy ontology and its application to news summarization. IEEE Transactions on Systems Man and Cybernetics Part B Cybernetics, 35(5), 859–880.

    Article  Google Scholar 

  • Lesser, V. R., & Corkill, D. D. (1988). The distributed vehicle monitoring testbed: A tool for investigating distributed problem solving networks, blackboard systems (pp. 353–386). Addison-Wesley.

  • Mirbakhsh, N., Didandeh, A., & Afsharchi, M. (2009). Incremental non-unanimous concept reformation through queried object classification. In Proc. 2009 IEEE/WIC/ACM international joint conference on Web intelligence and intelligent agent technology (Vol. 02, pp. 573–580).

  • Mirbakhsh, N., Didandeh, A., & Afsharchi, M. (2010). Concept learning games: The game of query and response. In Proc. 2010 IEEE/WIC/ACM international joint conference on Web intelligence and intelligent agent technology.

  • Mitchell, T. M. (1997). Machine learning. McGraw-Hill Science/Engineering/Math.

  • Osborne, M. J., & Rubinstein, A. (1995). A course in game theory. M.I.T. Press.

  • Packer, H., Payne, T., Gibbins, N., & Jennings, N. R. (2008). Evolving ontological knowledge bases through agent collaboration. In Proc. 6th European workshop on multi-agent systems.

  • Panait, L., & Luke, S. (2005). Cooperative multi-agent learning: The state of the art. Journal of Autonomous Agents and Multi-Agent Systems, 11(3), 387–434.

    Article  Google Scholar 

  • Parsons, S., & Wooldridge, M. (2000). Game theory and decision theory in multi-agent systems. Kluwer Academic Publishers.

  • Pitt, J., Kamara, L., Sergot, M., & Artikis, A. (2006). Voting in multi-agent systems. The Computer Journal, 49(2), 156–170.

    Article  Google Scholar 

  • Protégé Project (2006). Protégé 3.4.7. http://protege.stanford.edu/. Accessed July 2011.

  • Sen, S., & Kar, P. P. (2002). Sharing a concept. In Working notes of the AAAI spring symposium on collaborative learning agents.

  • Smith, R. G. (1977). The contract net: A formalism for the control of distributed problem solving. In Proc. 5th international joint conference on artificial intelligence (Vol. 1, pp. 472–472).

  • Stumme, G. (2001). Using ontologies and formal concept analysis for organizing business knowledge. In Proc. Referenzmodellierung 2001, Physica (pp. 163–174).

  • University of Michigan academic units (2005). List of offered courses, computer science. http://www.umich.edu/units.html/. Accessed on 30 January 2005.

  • van Diggelen, J., Beun, R. J., Dignum, F., van Eijk, R. M., & Meyer, J.-J. Ch. (2006). ANEMONE: An effective minimal ontology negotiation environment. In Proc. autonomous agents and multi agent systems 2006 (pp. 899–906).

  • van Diggelen, J., Beun, R. J., Dignum, F., van Eijk, R. M., & Meyer, J. J. (2007). Ontology negotiation: Goals, requirements and implementation. International Journal of Agent-Oriented Software Engineering, 1(1), 63–90.

    Article  Google Scholar 

  • Weibull, J. W. (1997). Evolutionary game theory. M.I.T. Press.

  • Williams, A. B. (2004). Learning to share meaning in a multi agent system. Journal of Autonomous Agents and Multi Agent Ssytems, 8(2), 165–193.

    Article  Google Scholar 

  • Wooldrige, M. (2009). An introduction to multiAgent system 2nd Edn. John Wiley and Sons.

  • Xianjia, W., & Weibing, L. (2009). Preference and evolution in the iterated prisoner’s dilemma. Acta Mathematica Scientia, 29B(2), 456–464.

    Article  Google Scholar 

  • Yang, Y., & Pedersen, J. O. (1997). A comparative study on feature selection in text categorization. In Proc. 14th international conference on machine learning (ICML-97) (pp. 412–420).

  • Zuber, V. S., & Faltings, B. (2006). Inferring user’s preferences using ontologies. In Proc. the 21st national conference on artificial intelligence (Vol. 2, pp. 1413–1418). AAAI Press.

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Nima Mirbakhsh.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Didandeh, A., Mirbakhsh, N. & Afsharchi, M. Concept learning games. Inf Syst Front 15, 653–676 (2013). https://doi.org/10.1007/s10796-012-9343-3

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10796-012-9343-3

Keywords

Navigation