Structured threshold policies for dynamic sensor scheduling—A partially observed Markov decision process approach

V Krishnamurthy, DV Djonin - IEEE Transactions on Signal …, 2007 - ieeexplore.ieee.org
V Krishnamurthy, DV Djonin
IEEE Transactions on Signal Processing, 2007ieeexplore.ieee.org
We consider the optimal sensor scheduling problem formulated as a partially observed
Markov decision process (POMDP). Due to operational constraints, at each time instant, the
scheduler can dynamically select one out of a finite number of sensors and record a noisy
measurement of an underlying Markov chain. The aim is to compute the optimal
measurement scheduling policy, so as to minimize a cost function comprising of estimation
errors and measurement costs. The formulation results in a nonstandard POMDP that is …
We consider the optimal sensor scheduling problem formulated as a partially observed Markov decision process (POMDP). Due to operational constraints, at each time instant, the scheduler can dynamically select one out of a finite number of sensors and record a noisy measurement of an underlying Markov chain. The aim is to compute the optimal measurement scheduling policy, so as to minimize a cost function comprising of estimation errors and measurement costs. The formulation results in a nonstandard POMDP that is nonlinear in the information state. We give sufficient conditions on the cost function, dynamics of the Markov chain and observation probabilities so that the optimal scheduling policy has a threshold structure with respect to a monotone likelihood ratio (MLR) ordering. As a result, the computational complexity of implementing the optimal scheduling policy is inexpensive. We then present stochastic approximation algorithms for estimating the best linear MLR order threshold policy.
ieeexplore.ieee.org