skip to main content
10.1145/1329125.1329182acmotherconferencesArticle/Chapter ViewAbstractPublication PagesaamasConference Proceedingsconference-collections
research-article

A reinforcement learning based distributed search algorithm for hierarchical peer-to-peer information retrieval systems

Published:14 May 2007Publication History

ABSTRACT

The dominant existing routing strategies employed in peer-to-peer(P2P) based information retrieval(IR) systems are similarity-based approaches. In these approaches, agents depend on the content similarity between incoming queries and their direct neighboring agents to direct the distributed search sessions. However, such a heuristic is myopic in that the neighboring agents may not be connected to more relevant agents. In this paper, an online reinforcement-learning based approach is developed to take advantage of the dynamic run-time characteristics of P2P IR systems as represented by information about past search sessions. Specifically, agents maintain estimates on the downstream agents' abilities to provide relevant documents for incoming queries. These estimates are updated gradually by learning from the feedback information returned from previous search sessions. Based on this information, the agents derive corresponding routing policies. Thereafter, these agents route the queries based on the learned policies and update the estimates based on the new routing policies. Experimental results demonstrate that the learning algorithm improves considerably the routing performance on two test collection sets that have been used in a variety of distributed IR studies.

References

  1. S. Abdallah and V. Lesser. Learning the task allocation game. In AAMAS '06: Proceedings of the fifth international joint conference on Autonomous agents and multiagent systems, pages 850--857, New York, NY, USA, 2006. ACM Press. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. J. A. Boyan and M. L. Littman. Packet routing in dynamically changing networks: A reinforcement learning approach. In Advances in Neural Information Processing Systems, volume 6, pages 671--678. Morgan Kaufmann Publishers, Inc., 1994.Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. J. C. French, A. L. Powell, J. P. Callan, C. L. Viles, T. Emmitt, K. J. Prey, and Y. Mou. Comparing the performance of database selection algorithms. In Research and Development in Information Retrieval, pages 238--245, 1999. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. B. Horling, R. Mailler, and V. Lesser. Farm: A scalable environment for multi-agent development and evaluation. In Advances in Software Engineering for Multi-Agent Systems, pages 220--237, Berlin, 2004. Springer-Verlag.Google ScholarGoogle Scholar
  5. M. Littman and J. Boyan. A distributed reinforcement learning scheme for network routing. In Proceedings of the International Workshop on Applications of Neural Networks to Telecommunications, 1993.Google ScholarGoogle Scholar
  6. J. Lu and J. Callan. Federated search of text-based digital libraries in hierarchical peer-to-peer networks. In In ECIR'05, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. J. Lu and J. Callan. User modeling for full-text federated search in peer-to-peer networks. In ACM SIGIR 2006. ACM Press, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. C. D. Manning and H. Schütze. Foundations of Statistical Natural Language Processing. The MIT Press, Cambridge, Massachusetts, 1999. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. C. R. Palmer and J. G. Steffan. Generating network topologies that obey power laws. In Proceedings of GLOBECOM '2000, November 2000.Google ScholarGoogle ScholarCross RefCross Ref
  10. K. Sripanidkulchai, B. Maggs, and H. Zhang. Efficient content location using interest-based locality in peer-topeer systems. In INFOCOM, 2003.Google ScholarGoogle ScholarCross RefCross Ref
  11. D. Subramanian, P. Druschel, and J. Chen. Ants and reinforcement learning: A case study in routing in dynamic networks. In In Proceedings of the Fifteenth International Joint Conference on Artificial Intelligence, pages 832--839, 1997. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. J. N. Tao and L. Weaver. A multi-agent, policy gradient approach to network routing. In In Proceedings of the Eighteenth International Conference on Machine Learning, 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. H. Zhang, W. B. Croft, B. Levine, and V. Lesser. A multi-agent approach for peer-to-peer information retrieval. In Proceedings of Third International Joint Conference on Autonomous Agents and Multi-Agent Systems, July 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. H. Zhang and V. Lesser. Multi-agent based peer-to-peer information retrieval systems with concurrent search sessions. In Proceedings of the Fifth International Joint Conference on Autonomous Agents and Multi-Agent Systems, May 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. H. Zhang and V. R. Lesser. A dynamically formed hierarchical agent organization for a distributed content sharing system. In 2004 IEEE/WIC/ACM International Conference on Intelligent Agent Technology (IAT 2004), 20--24 September 2004, Beijing, China, pages 169--175. IEEE Computer Society, 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. A reinforcement learning based distributed search algorithm for hierarchical peer-to-peer information retrieval systems

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Other conferences
      AAMAS '07: Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems
      May 2007
      1585 pages
      ISBN:9788190426275
      DOI:10.1145/1329125

      Copyright © 2007 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 14 May 2007

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article

      Acceptance Rates

      Overall Acceptance Rate1,155of5,036submissions,23%

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader