A Nearer Optimal and Faster Trained Value Iteration ADP for Discrete-Time Nonlinear Systems
Adaptive dynamic programming (ADP) is generally implemented using three neural networks: model network, action network, and critic network. In the conventional works of the value iteration ADP, the model network is initialized randomly and trained by the backpropagation algorithm, whose results are...
Main Authors: | Junping Hu, Gen Yang, Zhicheng Hou, Gong Zhang, Wenlin Yang, Weijun Wang |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2021-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9326299/ |
Similar Items
-
Poly(ADP-ribosyl)ation /
by: Burkle, Alexander
Published: (2006) -
Reversing ADP-ribosylation
by: Giuliana Katharina Moeller, et al.
Published: (2017-08-01) -
Event-Triggered Single-Network ADP for Zero-Sum Game of Unknown Nonlinear Systems with Constrained Input
by: Binbin Peng, et al.
Published: (2023-02-01) -
ARH Family of ADP-Ribose-Acceptor Hydrolases
by: Hiroko Ishiwata-Endo, et al.
Published: (2022-11-01) -
Uncovering the Invisible: Mono-ADP-ribosylation Moved into the Spotlight
by: Ann-Katrin Hopp, et al.
Published: (2021-03-01)