期刊文献+
共找到4篇文章
< 1 >
每页显示 20 50 100
具有模型和实际差异的非线性系统最优控制算法及其收敛性(英文) 被引量:2
1
作者 李俊民 邢科义 万百五 《控制理论与应用》 EI CAS CSCD 北大核心 1999年第3期57-61,共5页
针对模型和实际之间的差异,提出了一种基于时变线性二次型问题的动态系统优化和参数估计集成的算法,该算法能逼近实际问题最优解.给出了该算法收敛的一个充分条件,分析了它的最优性.仿真例子说明了该算法的有效性和实用性.
关键词 非线性系统最优控制 模型与实际差异 收敛性 最优性
全文增补中
Science Letters:A minimax optimal control strategy for uncertain quasi-Hamiltonian systems
2
作者 Yong WAN Zu-guang YIN Wei-qiu ZHU 《Journal of Zhejiang University-Science A(Applied Physics & Engineering)》 SCIE EI CAS CSCD 2008年第7期950-954,共5页
A minimax optimal control strategy for quasi-Hamiltonian systems with bounded parametric and/or external disturbances is proposed based on the stochastic averaging method and stochastic differential game. To conduct t... A minimax optimal control strategy for quasi-Hamiltonian systems with bounded parametric and/or external disturbances is proposed based on the stochastic averaging method and stochastic differential game. To conduct the system energy control, the partially averaged Ito stochastic differential equations for the energy processes are first derived by using the stochastic averaging method for quasi-Hamiltonian systems. Combining the above equations with an appropriate performance index, the proposed strategy is searching for an optimal worst-case controller by solving a stochastic differential game problem. The worst-case disturbances and the optimal controls are obtained by solving a Hamilton-Jacobi-Isaacs (HJI) equation. Numerical results for a controlled and stochastically excited DulTlng oscillator with uncertain disturbances exhibit the efficacy of the proposed control strategy. 展开更多
关键词 Nonlinear quasi-Hamiltonian system Minimax optimal control Stochastic excitation Uncertain disturbance Stochastic averaging Stochastic differential game
在线阅读 下载PDF
Bellman Equation for Optimal Processes with Nonlinear Multi-Parametric Binary Dynamic System
3
作者 Yakup H. Hacl Kemal Ozen 《Computer Technology and Application》 2012年第1期84-87,共4页
A process represented by nonlinear multi-parametric binary dynamic system is investigated in this work. This process is characterized by the pseudo Boolean objective functional. Since the transfer functions on the pro... A process represented by nonlinear multi-parametric binary dynamic system is investigated in this work. This process is characterized by the pseudo Boolean objective functional. Since the transfer functions on the process are Boolean functions, the optimal control problem related to the process can be solved by relating between the transfer functions and the objective functional. An analogue of Bellman function for the optimal control problem mentioned is defined and consequently suitable Bellman equation is constructed. 展开更多
关键词 Bellman equation bellman function galois field shift operator nonlinear multi-parametric binary dynamic system
在线阅读 下载PDF
A novel policy iteration based deterministic Q-learning for discrete-time nonlinear systems 被引量:8
4
作者 WEI QingLai LIU DeRong 《Science China Chemistry》 SCIE EI CAS CSCD 2015年第12期143-157,共15页
In this paper, a novel iterative Q-learning algorithm, called "policy iteration based deterministic Qlearning algorithm", is developed to solve the optimal control problems for discrete-time deterministic no... In this paper, a novel iterative Q-learning algorithm, called "policy iteration based deterministic Qlearning algorithm", is developed to solve the optimal control problems for discrete-time deterministic nonlinear systems. The idea is to use an iterative adaptive dynamic programming(ADP) technique to construct the iterative control law which optimizes the iterative Q function. When the optimal Q function is obtained, the optimal control law can be achieved by directly minimizing the optimal Q function, where the mathematical model of the system is not necessary. Convergence property is analyzed to show that the iterative Q function is monotonically non-increasing and converges to the solution of the optimality equation. It is also proven that any of the iterative control laws is a stable control law. Neural networks are employed to implement the policy iteration based deterministic Q-learning algorithm, by approximating the iterative Q function and the iterative control law, respectively. Finally, two simulation examples are presented to illustrate the performance of the developed algorithm. 展开更多
关键词 adaptive critic designs adaptive dynamic programming approximate dynamic programming Q-LEARNING policy iteration neural networks nonlinear systems optimal control
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部