Abstract:
In the context of Markov Decision Processes (MDPs), the framework of forward-backward probability propagation on factor graphs has proven to be useful for finding optimal...Show MoreMetadata
Abstract:
In the context of Markov Decision Processes (MDPs), the framework of forward-backward probability propagation on factor graphs has proven to be useful for finding optimal policies. However, in cases involving vector rewards, there is a need to evaluate a trade-off among constituent objectives. In this work, assuming multiple rewards, we show how to use the framework of belief propagation for dynamically generating the Pareto front and propagating it as a forward flow distribution. The idea is applied to path planning on discrete 1D and 2D grids where different sets of states have vector rewards in the form of priors.
Published in: 2023 IEEE 33rd International Workshop on Machine Learning for Signal Processing (MLSP)
Date of Conference: 17-20 September 2023
Date Added to IEEE Xplore: 23 October 2023
ISBN Information: