Skip to main content

Part of the book series: Studies in Computational Intelligence ((SCI,volume 80))

Summary

This chapter discusses transfer learning, which is one practical application of rule extraction. In transfer learning, information from one learning experience is applied to speed up learning in a related task. The chapter describes several techniques for transfer learning in SVM-basedreinforcement learning, and shows results from a case study.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. R. Andrews, J. Diederich, and A. Tickle. A survey and critique of techniques for extracting rules from trained artificial neural networks. In Knowledge Based Systems, 1995.

    Google Scholar 

  2. J. Clouse and P. Utgoff. A teaching method for reinforcement learning. In Proceedings of the 9th International Conference on Machine Learning, 1992.

    Google Scholar 

  3. N. Cristianini and J. Shawe-Taylor. An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods. Cambridge University Press, 2000.

    Google Scholar 

  4. K. Driessens and S. Dzeroski. Integrating experimentation and guidance in relational reinforcement learning. In Proceedings of the 19th International Conference on Machine Learning, 2002.

    Google Scholar 

  5. G. Kuhlmann, P. Stone, R. Mooney, and J. Shavlik. Guiding a reinforcement learner with natural language advice: Initial results in RoboCup soccer. In AAAI Workshop on Supervisory Control of Learning and Adaptive Systems, 2004.

    Google Scholar 

  6. R. Maclin, J. Shavlik, L. Torrey, and T. Walker. Knowledge-based support vector regression for reinforcement learning. In IJCAI Workshop on Reasoning, Representation, and Learning in Computer Games, 2005.

    Google Scholar 

  7. R. Maclin, J. Shavlik, L. Torrey, T. Walker, and E. Wild. Giving advice about preferred actions to reinforcement learners via knowledge-based kernel regression. In Proceedings of the 20th National Conference on Artificial Intelligence, 2005.

    Google Scholar 

  8. O. Mangasarian, J. Shavlik, and E. Wild. Knowledge-based kernel approximation. Journal of Machine Learning Research 5, pages 1127–1141, 2004.

    MathSciNet  Google Scholar 

  9. S. Muggleton and L. De Raedt. Inductive logic programming: Theory and methods. Journal of Logic Programming 19,20, pages 629–679, 1994.

    Article  MathSciNet  Google Scholar 

  10. I. Noda, H. Matsubara, K. Hiraki, and I. Frank. Soccer server: A tool for research on multiagent systems. Applied Artificial Intelligence, 12:233–250, 1998.

    Article  Google Scholar 

  11. V. Soni and S. Singh. Using homomorphisms to transfer options across continuous reinforcement learning domains. In Proceedings of the 21st National Conference on Artificial Intelligence, 2006.

    Google Scholar 

  12. A. Srinivasan. The Aleph Manual, 2001.

    Google Scholar 

  13. P. Stone and R. Sutton. Scaling reinforcement learning toward RoboCup soccer. In Proceedings of the 18th International Conference on Machine Learning, 2001.

    Google Scholar 

  14. R. Sutton and A. Barto. Reinforcement Learning: An Introduction. MIT Press, 1998.

    Google Scholar 

  15. M. Taylor, P. Stone, and Y. Liu. Value functions for RL-based behavior transfer: A comparative study. In Proceedings of the 20th National Conference on Artificial Intelligence, 2005.

    Google Scholar 

  16. L. Torrey, J. Shavlik, T. Walker, and R. Maclin. Advice-based transfer in reinforcement learning. Technical Report TR06-2, Machine Learning Research Group, U. Wisconsin-Madison, 2006.

    Google Scholar 

  17. L. Torrey, J. Shavlik, T. Walker, and R. Maclin. Relational skill transfer via advice taking. In Proceedings of the 17th European Conference on Machine Learning, 2006.

    Google Scholar 

  18. L. Torrey, T. Walker, J. Shavlik, and R. Maclin. Using advice to transfer knowledge acquired in one reinforcement learning task to another. In Proceedings of the 16th European Conference on Machine Learning, 2005.

    Google Scholar 

  19. C. Watkins. Learning from delayed rewards. Technical Report PhD Thesis, University of Cambridge, Psychology Dept., 1989.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2008 Springer-Verlag Berlin Heidelberg

About this chapter

Cite this chapter

Torrey, L., Shavlik, J., Walker, T., Maclin, R. (2008). Rule Extraction for Transfer Learning. In: Diederich, J. (eds) Rule Extraction from Support Vector Machines. Studies in Computational Intelligence, vol 80. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-75390-2_3

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-75390-2_3

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-75389-6

  • Online ISBN: 978-3-540-75390-2

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics