Skip to main content
Log in

Statistical Relational Artificial Intelligence: From Distributions through Actions to Optimization

  • Technical Contribution
  • Published:
KI - Künstliche Intelligenz Aims and scope Submit manuscript

Abstract

Statistical Relational AI—the science and engineering of making intelligent machines acting in noisy worlds composed of objects and relations among the objects—is currently motivating a lot of new AI research and has tremendous theoretical and practical implications. Theoretically, combining logic and probability in a unified representation and building general-purpose reasoning tools for it has been the dream of AI, dating back to the late 1980s. Practically, successful statistical relational AI tools enable new applications in several large, complex real-world domains including those involving big data, natural text, social networks, the web, medicine and robotics, among others. Such domains are often characterized by rich relational structure and large amounts of uncertainty. Logic helps to faithfully model the former while probability helps to effectively manage the latter. Our intention here is to give a brief (and necessarily incomplete) overview and invitation to the emerging field of Statistical Relational AI from the perspective of acting optimally and learning to act.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

Notes

  1. Here, at each time step, a system is in some state s, and the agent may choose any action a that is available in state s. Then, depending on a, the system moves stochastically into state \(s^\prime\), and the agent receives a reward \(R_a(s,s^\prime )\).

  2. If there are more than one box in Paris, we still get only a reward of 10 due to the existential quantifier of b. That is, the formulas in the reward can be viewed as logical queries that are either true or false. If they are true, we get the reward.

  3. Likewise, for solving FOL-MPDs, Sanner and Boutilier [46] used approximate linear programming with relational constraints represented via case statements. This approach scaled to problems of previously prohibitive size by avoiding grounding and is indeed close in spirit to relational linear programming.

References

  1. Abdo N, Kretzschmar H, Spinello L, Stachniss C (2013) Learning manipulation actions from a few demonstrations. In: 2013 IEEE international conference on robotics and automation (ICRA), pp 1268–1275

  2. Ahmadi B, Kersting K, Mladenov M, Natarajan S (2013) Exploiting symmetries for scaling loopy belief propagation and relational training. Mach Learn 92(1):91–132

    Article  MathSciNet  MATH  Google Scholar 

  3. Bacchus F, Halpern JY, Levesque HJ (1999) Reasoning about noisy sensors and effectors in the situation calculus. Artificial Intelligence 111(1–2):171–208 (1999). URL:http://www.lpaig.uwaterloo.ca/fbacchus/on-line.html

  4. Beetz M, Jain D, Mösenlechner L, Tenorth M (2010) Towards performing everyday manipulation activities. Robot Auton Syst 58(9):1085–1095

    Article  Google Scholar 

  5. Bellman RE (1957) Dynamic programming. Princeton University Press, Princeton

    MATH  Google Scholar 

  6. Boutilier C, Reiter R, Price B (2001) Symbolic dynamic programming for first-order MDPs. In: Nebel B (ed) Seventeenth international joint conference on artificial intelligence (IJCAI-01). Morgan Kaufmann, Seattle, USA, pp 690–700

    Google Scholar 

  7. Van den Broeck G, Thon I, van Otterlo M, De Raedt L (2010) DTProbLog: A decision-theoretic probabilistic prolog. In: Proceedings of the AAAI conference on artificial intelligence (AAAI 2010)

  8. Driessens K, Dzeroski S (2002) Integrating experimentation and guidance in relational reinforcement learning. In: Proceedings of ICML, pp 115–122

  9. Driessens K, Dzeroski S (2005) Combining model-based and instance-based learning for first-order regression, pp 193–200

  10. Driessens K, Ramon J (2003) Relational instance based regression for relational reinforcement learning, pp 123–130

  11. Driessens K, Ramon J, Gärtner T (2006) Graph kernels and Gaussian processes for relational reinforcement learning. Mach Learn J 64(1–3):91–119

    Article  MATH  Google Scholar 

  12. Džeroski S, De Raedt L, Driessens K (2001) Relational reinforcement learning. Mach Learn J 43(1/2):7–52

    Article  MATH  Google Scholar 

  13. Fern A, Yoon S, Givan R (2003) Approximate policy iteration with a policy language bias. In: NIPS-2003. Vancouver

  14. Fern A, Yoon S, Givan R (2006) Approximate policy iteration with a policy language bias: Solving relational markov decision processes. J Artif Intell Res (JAIR) 25:75–118

    MathSciNet  MATH  Google Scholar 

  15. Fikes RE, Nilsson NJ (1971) STRIPS: a new approach to the application of theorem proving to problem solving. Artif Intell 2(3–4):189–208

    Article  MATH  Google Scholar 

  16. Geffner H (2014) Artificial intelligence: from programs to solvers. AI Commun. 27(1):45–51

    MathSciNet  Google Scholar 

  17. Getoor L (2007) In: Taskar B (eds.) An introduction to statistical relational learning. MIT Press, Cambridge

  18. Gretton C, Thiebaux S (2004) Exploiting first-order regression in inductive policy selection. In: UAI-04, pp 217–225

  19. Hölldobler S, Skvortsova O (2004) A logic-based approach to dynamic programming. In: In AAAI-04 Workshop on Learning and planning in MDPs, pp 31–36. Menlo Park, CA

  20. Joshi S, Kersting K, Khardon R (2010) Self-taught decision theoretic planning with first order decision diagrams. In: Proceedings of ICAPS-10

  21. Kaelbling LP, Littman ML, Cassandra AR (1998) Planning and acting in partially observable stochastic domains. Artif Intell J 101(1–2):99–134

    Article  MathSciNet  MATH  Google Scholar 

  22. Karabaev E, Skvortsova O (2005) A heuristic search algorithm for solving first-order MDPs. In: UAI-2005, pp 292–299

  23. Kersting K (2012) Lifted probabilistic inference. In: Proceedings of the 20th European conference on artificial intelligence (ECAI), pp 33–38. (Invited Talk at the Frontiers of AI Track)

  24. Kersting K, Driessens K (2008) Non-parametric policy gradients: A unified treatment of propositional and relational domains. In: McCallum SRA (ed) Proceedings of the 25th international conference on machine learning (ICML 2008). Helsinki, Finland

    Google Scholar 

  25. Kersting K, Mladenov M, Tokmakov P (2015) Relational linear programming. Artif Intell J. (Accepted for publication)

  26. Kersting K, van Otterlo M, de Raedt L (2004) Bellman goes relational. In: ICML-04. ACM Press, Banff, Alberta, Canada. doi:10.1145/1015330.1015401

  27. Kimmig A, Mihalkova L, Getoor L (2015) Lifted graphical models: a survey. Mach Learn 99(1):1–45

    Article  MathSciNet  Google Scholar 

  28. Koller D, Friedman N (2009) Probabilistic graphical models—principles and techniques. MIT Press, Cambridge

  29. Kulick J, Toussaint M, Lang T, Lopes M (2013) Active learning for teaching a robot grounded relational symbols. In: Proceedings of the 23rd international joint conference on artificial intelligence (IJCAI)

  30. Lang T, Toussaint M (2009) Relevance grounding for planning in relational domains. In Buntine WL, Grobelnik M, Mladenic G, Shawe-Taylo J (eds) Proceedings of the european conference of machine learning and knowledge discovery in databases (ECML/PKDD), Part I. Springer, pp 736–751

  31. Lang T, Toussaint M (2010) Planning with noisy probabilistic relational rules. J Artif Intell Res 39:1–49

    MATH  Google Scholar 

  32. Lang T, Toussaint M, Kersting K (2010) Exploration in relational worlds. In: Balcazar J, Bonchi F, Gionis A, Sebag M (eds) European conference on machine learning and principles and practice of knowledge discovery in databases (ECML PKDD-10). Springer, Barcelona

    Google Scholar 

  33. Lang T, Toussaint M, Kersting K (2012) Exploration in relational domains for model-based reinforcement learning. J Mach Learn Res 13:3691–3734

    MathSciNet  Google Scholar 

  34. Natarajan S, Joshi S, Tadepalli P, Kersting K, Shavlik J (2011) Imitation learning in relational domains: a functional-gradient boosting approach. In: Proceedings of the 22nd International Joint Conference on Artificial Intelligence (IJCAI), pp 1414–1420

  35. Nath A, Richardson M (2012) Counting-mlns: Learning relational structure for decision making. In: Proceedings of the twenty-sixth AAAI conference on artificial intelligence

  36. Ngo V, Toussaint M (2014) Model-based relational RL when object existence is partially observable. In: Proceedings of the 31th international conference on machine learning (ICML), pp 559–567

  37. Orthey A, Toussaint M, Jetchev N (2013) Optimizing motion primitives to make symbolic models more predictive. In: 2013 IEEE international conference on robotics and automation (ICRA), pp 2868–2873

  38. Pasula H, Zettlemoyer L, Pack Kaelbling L (2007)Learning symbolic models of stochastic domains. J Artif Intell Res 29:309–352

  39. Pearl J (1991) Reasoning in intelligent systems: networks of plausible Inference, 2nd edn. Morgan Kaufmann, San Francisco, CA, USA

  40. Poole D (1997) The independent choice logic for modelling multiple agents under uncertainty. Artif Intell J 94:7–56

    Article  MathSciNet  MATH  Google Scholar 

  41. Poole D (1998) Decision theory, the situation calculus and conditional plans. Electron Trans Artif Intell 2(1–2). URL:http://www.etaij.org

  42. Powell WB (2010) Merging ai and or to solve high-dimensional stochastic optimization problems using approximate dynamic programming. INFORMS INFORMS J Comput 22(1):2–17

    Article  MathSciNet  MATH  Google Scholar 

  43. Ramon J, Driessens K, Croonenborghs T (2007) Transfer learning in reinforcement learning problems through partial policy recycling. In: Proceedings of ECML, pp 699–707

  44. Sanner S, Boutilier C (2007) Approximate solution techniques for factored first-order MDPs. In: Proceedings of ICAPS

  45. Sanner S, Boutilier C (2009) Practical solution techniques for first order mdps. AIJ 173:748–788

    MathSciNet  MATH  Google Scholar 

  46. Sanner S, Boutilier C (2009) Practical solution techniques for first-order MDPs. Artif Intell J 173(5–6):748–788

    Article  MathSciNet  MATH  Google Scholar 

  47. Sanner S, Kersting K (2010) Symbolic dynamic programming. In: C. Sammut, G. Webb (eds.) Encyclopedia of machine learning, p 946–954. Springer

  48. Sanner S, Kersting K (2010) Symbolic dynamic programming for first-order POMDPs. In: Proc. AAAI-2010

  49. Sarjant S, Pfahringer B, Driessens K, Smith T (2011) Using the online cross-entropy method to learn relational policies for playing different games. In: 2011 IEEE conference on computational intelligence and games (CIG), pp 182–189

  50. Sutton RS, Barto AG (1998) Reinforcement learning: an Introduction. MIT Press, Canbridge, MA

    Google Scholar 

  51. Tadepalli P, Givan R, Driessens K (2004) Relational reinforcement learning: an overview. In: Proc. ICML workshop on relational reinforcement learning

  52. Tenorth M, Klank U, Pangercic D, Beetz M (2011) Web-enabled robots—robots that use the web as an information resource. IEEE robotics and automation magazine p 18

  53. Van Haaren J, Van den Broeck G, Meert W, Davis J (2014) Tractable learning of liftable Markov logic networks. In: Proceedings of the ICML-14 workshop on learning tractable probabilistic models (LTPM)

  54. van Otterlo M (2009) The logic of adaptive behavior−knowledge representation and algorithms for adaptive sequential decision making under uncertainty in first-order and relational domains. IOS Press

  55. Wang C, Joshi S, Khardon R (2008) First order decision diagrams for relational mdps. JAIR 31:431–472

    MathSciNet  MATH  Google Scholar 

  56. Wang C, Khardon R (2010) Relational partially observable MDPs. In: Proc. AAAI-2010

  57. Zamani Z, Sanner S, Poupart P, Kersting K (2012) Symbolic dynamic programming for continuous state and observation pomdps. In: Proceedings of the 26th annual conference on neural information processing systems (NIPS), pp 1403–1411

Download references

Acknowledgments

The authors thank the anonymous reviewers for their feedback. They are also grateful to all the people who contributed to the development of Statistical Relational AI, in particular to Scott Sanner and David Poole for previous and current joint efforts in writing introductions and overviews on statistical relational AI and symbolic dynamic programming, parts of which grew into the present paper. KK also likes to acknowledge the supported by by the German Science Foundation (DFG), KE 1686/2-1, within the SPP 1527 “Autonomous Learning”, and SN the support of Army Research Office (ARO) grant number W911NF-13-1-0432 under the Young Investigator Program.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Kristian Kersting.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Kersting, K., Natarajan, S. Statistical Relational Artificial Intelligence: From Distributions through Actions to Optimization. Künstl Intell 29, 363–368 (2015). https://doi.org/10.1007/s13218-015-0386-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s13218-015-0386-8

Keywords

Navigation