Abstract
In probabilistic accounts of belief change, traditionally Bayesian conditioning is employed when the received information is consistent with the current knowledge, and imaging is used otherwise. It is well recognised that imaging can be used even if the received information is consistent with the current knowledge. Imaging assumes, inter alia, a relational measure of similarity among worlds. In a recent work, Rens and Meyer have argued that when, in light of new evidence, we no longer consider a world \(\omega \) to be a serious possibility, worlds more similar to it should be considered relatively less plausible, and hence more dissimilar (distant) a world is from \(\omega \), the larger should be its share in the original probability mass of \(\omega \). In this paper we argue that this approach leads to results that revolt against our causal intuition, and propose a converse account where a larger share of \(\omega \)’s mass move to worlds that are more similar (closer) to it instead.
This research has been partially supported by the Australian Research Council (ARC), Discovery Project: DP150104133.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Assuming such a unique closest world exists. Short of it, an appropriate distribution mechanism should be employed.
- 2.
Recall the standard refrain of the politicians, I don’t answer hypothetical questions.
- 3.
The distance between different A-worlds, or between different \(\overline{A}\)-worlds is not shown since they will not be used in the calculation.
- 4.
The more distant/different a world is from a target world, the higher is its relative weight.
- 5.
We need not worry about adding the new probabilities of relevant \(\overline{A}\)-worlds since they are all zero.
- 6.
Arguably the use of similarity in common parlance is non-symmetric. For instance, if John is non-violent, we would say John is like Gandhi. But saying Gandhi is like John would mean a very different thing. Capturing such asymmetry in our simple framework may not be quite feasible.
References
Alchourrón, C.E., Gärdenfors, P., Makinson, D.: On the logic of theory change: partial meet contraction and revision functions. J. Symb. Log. 50(2), 510–530 (1985)
Baral, C., Tran, N., Tuan, L.: Reasoning about actions in a probabilistic setting. In: Proceedings of the AAAI, pp. 507–512 (2002)
Chhogyal, K., Nayak, A., Schwitter, R., Sattar, A.: Probabilistic belief revision via imaging. In: Pham, D.-N., Park, S.-B. (eds.) PRICAI 2014. LNCS(LNAI), vol. 8862, pp. 694–707. Springer, Heidelberg (2014). doi:10.1007/978-3-319-13560-1_55
Gärdenfors, P.: Knowledge in Flux. MIT Press, Cambridge (1988)
Hansson, S.O.: Knowledge-level analysis of belief base operations. Artif. Intell. 82(1–2), 215–235 (1996)
Iocchi, L., Lukasiewicz, T., Nardi, D., Rosati, R.: Reasoning about actions with sensing under qualitative and probabilistic uncertainty. ACM Trans. Comput. Log. 10(1) (2009). doi:10.1145/1459010.1459015
Katsuno, H., Mendelzon, A.O.: On the difference between updating a knowledge base and revising it. In: Proceedings of the KR, pp. 387–394 (1991)
Lehmann, D.J., Magidor, M., Schlechta, K.: Distance semantics for belief revision. J. Symb. Log. 66(1), 295–317 (2001)
Lewis, D.: Probabilities of conditionals and conditional probabilities. Philos. Rev. 85(3), 297–315 (1976)
Nayak, A.C.: Iterated belief change based on epistemic entrenchment. Erkenntnis 41, 353–390 (1994)
Nayak, A., Goebel, R., Orgun, M., Pham, T.: Taking Levi Identity seriously: a plea for iterated belief contraction. In: Lehner, F., Fteimi, N. (eds.) KSEM 2016. LNCS(LNAI), vol. 9983, pp. 305–317. Springer, Heidelberg (2006). doi:10.1007/11811220_26
Pearl, J.: Probabilistic Reasoning in Intelligent Systems - Networks of Plausible Inference. Morgan Kaufmann, Burlington (1989)
Rens, G.B., Meyer, T.: A new approach to probabilistic belief change. In: Proceedings of the FLAIRS, pp. 581–587 (2015)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing AG
About this paper
Cite this paper
Mishra, S., Nayak, A. (2016). Causal Basis for Probabilistic Belief Change: Distance vs. Closeness. In: Sombattheera, C., Stolzenburg, F., Lin, F., Nayak, A. (eds) Multi-disciplinary Trends in Artificial Intelligence. MIWAI 2016. Lecture Notes in Computer Science(), vol 10053. Springer, Cham. https://doi.org/10.1007/978-3-319-49397-8_10
Download citation
DOI: https://doi.org/10.1007/978-3-319-49397-8_10
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-49396-1
Online ISBN: 978-3-319-49397-8
eBook Packages: Computer ScienceComputer Science (R0)