Operational Absolutely Optimal Dynamic Control of the Stochastic Differential Plant’s State by Its Output

Cover Page

Cite item

Full Text

Open Access Open Access
Restricted Access Access granted
Restricted Access Subscription Access

Abstract

The problem of synthesizing the average-optimal control law for a dynamic plant subject to random disturbances, if its state variables are measured partially or with random errors, is considered. Using the method of a posteriori sufficient coordinates (SCs), the complexity of constructing the well-known interval-optimal Mortensen controller is described and a much simpler algorithm for finding its operational-optimal analog is obtained. The new controller does not require the solution of the corresponding Bellman equation in inverse time, since it is optimal in the sense of a time-varying criterion. This makes it possible to disregard information about the future behavior of the object and reduces the procedure for finding the dependence of a control on sufficient coordinates to direct-time integration of the Fokker–Planck–Kolmogorov equation and to solving a problem of parametric nonlinear programming. The application of the obtained algorithm is demonstrated by the example of a linear-quadratic-Gaussian problem, as a result of which a new operational version of the well-known separation theorem is formulated. It represents a stochastic control device as a combination of a linear Kalman–Bucy filter and a linear operational-optimal positional controller. The latter differs from the traditional interval-optimal controller by the well-known gain and does not require the solution of the corresponding matrix Riccati equation in inverse time

About the authors

E. A. Rudenko

Moscow Aviation Institute (National Research University), 125080, Moscow, Russia

Author for correspondence.
Email: rudenkoevg@yandex.ru
Россия, Москва

References

  1. Стратонович Р.Л. К теории оптимального управления. Достаточные координаты // АиТ. 1962. № 7. С. 910–917.
  2. Mortensen R.E. Stochastic Optimal Control with Noisy Observations // Int. J. Control. 1966. V. 4. № 5. P. 455–466.
  3. Davis M.H.A., Varaiya P.P. Dynamic Programming Conditions for Partially Observable Stochastic Systems // SIAM J. Control. 1973. V. 11. № 2. P. 226–262.
  4. Параев Ю.И. Введение в статистическую динамику процессов управления и фильтрации. М.: Сов. радио, 1976.
  5. Benes V.E., Karatzas I. On the Relation of Zakai’s and Mortensen’s Equations // SIAM J. Control and Optimization. 1983. V. 21. № 3. P. 472–489.
  6. Bensoussan A. Stochastic Control of Partially Observable Systems. Cambridge: Cambridge University Press, 1992.
  7. Руденко Е.А. Оперативно-оптимальный конечномерный динамический регулятор состояния стохастического дифференциального объекта по его выходу. I. Общий нелинейный случай // Изв. РАН. ТиСУ. 2022. № 5. С. 23–39.
  8. Wonham W.M. On the Separation Theorem of Stochastic Control // SIAM J. Control. 1968. V. 6. № 2. P. 312–326.
  9. Верба В.С., Меркулов В.И., Руденко Е.А. Линейно-кубическое локально-оптимальное управление линейными системами и его применение для наведения летательных аппаратов // Изв. РАН. ТиСУ. 2020. № 5. С. 129–141.
  10. Пугачев В.С., Синицын И.Н. Стохастические дифференциальные системы. Анализ и фильтрация. М.: Наука, 1985.
  11. Синицын И.Н. Фильтры Калмана и Пугачева. М.: Логос, 2007.
  12. Пантелеев А.В., Руденко Е.А., Бортаковский А.С. Нелинейные системы управления: описание, анализ и синтез. М.: Вузовская книга, 2008.
  13. Руденко Е.А. Оптимальная структура непрерывного нелинейного фильтра Пугачева пониженного порядка // Изв. РАН. ТиСУ. 2013. № 6. С. 25–51.
  14. Ширяев А.Н. Вероятность. М.: Наука, 1980.
  15. Браммер К., Зиффлинг Г. Фильтр Калмана–Бьюси / Пер. с англ. М.: Наука, 1982.


Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.

This website uses cookies

You consent to our cookies if you continue to use our website.

About Cookies