Discrete-time markovian-jump linear quadratic optimal control
- 1 January 1986
- journal article
- research article
- Published by Taylor & Francis in International Journal of Control
- Vol. 43 (1) , 213-231
- https://doi.org/10.1080/00207178608933459
Abstract
This paper is concerned with the optimal control of discrete-time linear systems that possess randomly jumping parameters described by finite-state Markov processes. For problems having quadratic costs and perfect observations, the optimal control laws and expected costs-to-go can be precomputed from a set of coupled Riccati-like matrix difference equations. Necessary and sufficient conditions are derived for the existence of optimal constant control laws which stabilize the controlled system as the time horizon becomes infinite, with finite optimal expected cost.Keywords
This publication has 7 references indexed in Scilit:
- Approximations for functionals and optimal control problems on jump diffusion processesJournal of Mathematical Analysis and Applications, 1978
- Feedback control of a class of linear discrete systems with jump parameters and quadratic cost criteria †International Journal of Control, 1975
- Dynamic Programming and Minimum Principles for Systems with Jump Markov DisturbancesSIAM Journal on Control, 1975
- A computational algorithm for design of regulators for linear jump parameter systemsIEEE Transactions on Automatic Control, 1974
- Feedback regulators for jump parameter systems with state and control dependent transition ratesIEEE Transactions on Automatic Control, 1973
- Bayes and minimax controllers for a linear system with stochastic jump parametersIEEE Transactions on Automatic Control, 1971
- Feedback control of a class of linear systems with jump parametersIEEE Transactions on Automatic Control, 1969