Abstract:
We study the problem of zero-delay coding of a Markov source over a noisy channel with feedback. Building and generalizing prior work, we first formulate the problem as a...Show MoreMetadata
Abstract:
We study the problem of zero-delay coding of a Markov source over a noisy channel with feedback. Building and generalizing prior work, we first formulate the problem as a Markov decision process (MDP) where the state is a probability measure valued predictor along with a finite memory of channel outputs and quantizers. We then approximate this state by marginalizing over all possible predictors, so that our policies only use the finite-memory term to encode the source. Under an appropriate notion of predictor stability, we show that such policies are near-optimal for the zero-delay coding problem as the memory length increases. We also give sufficient conditions for predictor stability to hold, and present a reinforcement learning algorithm and establish its convergence to compute near-optimal finite-memory policies. These theoretical results are supported by simulations.
Published in: 2024 American Control Conference (ACC)
Date of Conference: 10-12 July 2024
Date Added to IEEE Xplore: 05 September 2024
ISBN Information: