Adaptive Dynamic Programming: Single and Multiple Controllers

Adaptive Dynamic Programming: Single and Multiple Controllers
Author: Ruizhuo Song
Publisher: Springer
Total Pages: 278
Release: 2018-12-28
Genre: Technology & Engineering
ISBN: 9811317127

This book presents a class of novel optimal control methods and games schemes based on adaptive dynamic programming techniques. For systems with one control input, the ADP-based optimal control is designed for different objectives, while for systems with multi-players, the optimal control inputs are proposed based on games. In order to verify the effectiveness of the proposed methods, the book analyzes the properties of the adaptive dynamic programming methods, including convergence of the iterative value functions and the stability of the system under the iterative control laws. Further, to substantiate the mathematical analysis, it presents various application examples, which provide reference to real-world practices.

Advances in Neural Computation, Machine Learning, and Cognitive Research VII

Advances in Neural Computation, Machine Learning, and Cognitive Research VII
Author: Boris Kryzhanovsky
Publisher: Springer Nature
Total Pages: 505
Release: 2023-11-12
Genre: Technology & Engineering
ISBN: 3031448650

This book describes new theories and applications of artificial neural networks, with a special focus on answering questions in neuroscience, biology and biophysics and cognitive research. It covers a wide range of methods and technologies, including deep neural networks, large-scale neural models, brain–computer interface, signal processing methods, as well as models of perception, studies on emotion recognition, self-organization and many more. The book includes both selected and invited papers presented at the XXV International Conference on Neuroinformatics, held on October 23-27, 2023, in Moscow, Russia.

Adaptive Dynamic Programming

Adaptive Dynamic Programming
Author: Jiayue Sun
Publisher: Springer Nature
Total Pages: 144
Release: 2023-10-14
Genre: Technology & Engineering
ISBN: 9819959292

This open access book focuses on the practical application of Adaptive Dynamic Programming (ADP) in chemotherapy drug delivery, taking into account clinical variables and real-time data. ADP's ability to adapt to changing conditions and make optimal decisions in complex and uncertain situations makes it a valuable tool in addressing pressing challenges in healthcare and other fields. As optimization technology evolves, we can expect to see even more sophisticated and powerful solutions emerge.

Adaptive Dynamic Programming with Applications in Optimal Control

Adaptive Dynamic Programming with Applications in Optimal Control
Author: Derong Liu
Publisher: Springer
Total Pages: 609
Release: 2017-01-04
Genre: Technology & Engineering
ISBN: 3319508156

This book covers the most recent developments in adaptive dynamic programming (ADP). The text begins with a thorough background review of ADP making sure that readers are sufficiently familiar with the fundamentals. In the core of the book, the authors address first discrete- and then continuous-time systems. Coverage of discrete-time systems starts with a more general form of value iteration to demonstrate its convergence, optimality, and stability with complete and thorough theoretical analysis. A more realistic form of value iteration is studied where value function approximations are assumed to have finite errors. Adaptive Dynamic Programming also details another avenue of the ADP approach: policy iteration. Both basic and generalized forms of policy-iteration-based ADP are studied with complete and thorough theoretical analysis in terms of convergence, optimality, stability, and error bounds. Among continuous-time systems, the control of affine and nonaffine nonlinear systems is studied using the ADP approach which is then extended to other branches of control theory including decentralized control, robust and guaranteed cost control, and game theory. In the last part of the book the real-world significance of ADP theory is presented, focusing on three application examples developed from the authors’ work: • renewable energy scheduling for smart power grids;• coal gasification processes; and• water–gas shift reactions. Researchers studying intelligent control methods and practitioners looking to apply them in the chemical-process and power-supply industries will find much to interest them in this thorough treatment of an advanced approach to control.

Adaptive Dynamic Programming for Control

Adaptive Dynamic Programming for Control
Author: Huaguang Zhang
Publisher: Springer Science & Business Media
Total Pages: 432
Release: 2012-12-14
Genre: Technology & Engineering
ISBN: 144714757X

There are many methods of stable controller design for nonlinear systems. In seeking to go beyond the minimum requirement of stability, Adaptive Dynamic Programming in Discrete Time approaches the challenging topic of optimal control for nonlinear systems using the tools of adaptive dynamic programming (ADP). The range of systems treated is extensive; affine, switched, singularly perturbed and time-delay nonlinear systems are discussed as are the uses of neural networks and techniques of value and policy iteration. The text features three main aspects of ADP in which the methods proposed for stabilization and for tracking and games benefit from the incorporation of optimal control methods: • infinite-horizon control for which the difficulty of solving partial differential Hamilton–Jacobi–Bellman equations directly is overcome, and proof provided that the iterative value function updating sequence converges to the infimum of all the value functions obtained by admissible control law sequences; • finite-horizon control, implemented in discrete-time nonlinear systems showing the reader how to obtain suboptimal control solutions within a fixed number of control steps and with results more easily applied in real systems than those usually gained from infinite-horizon control; • nonlinear games for which a pair of mixed optimal policies are derived for solving games both when the saddle point does not exist, and, when it does, avoiding the existence conditions of the saddle point. Non-zero-sum games are studied in the context of a single network scheme in which policies are obtained guaranteeing system stability and minimizing the individual performance function yielding a Nash equilibrium. In order to make the coverage suitable for the student as well as for the expert reader, Adaptive Dynamic Programming in Discrete Time: • establishes the fundamental theory involved clearly with each chapter devoted to a clearly identifiable control paradigm; • demonstrates convergence proofs of the ADP algorithms to deepen understanding of the derivation of stability and convergence with the iterative computational methods used; and • shows how ADP methods can be put to use both in simulation and in real applications. This text will be of considerable interest to researchers interested in optimal control and its applications in operations research, applied mathematics computational intelligence and engineering. Graduate students working in control and operations research will also find the ideas presented here to be a source of powerful methods for furthering their study.

Robust Adaptive Dynamic Programming

Robust Adaptive Dynamic Programming
Author: Yu Jiang
Publisher: John Wiley & Sons
Total Pages: 220
Release: 2017-04-13
Genre: Science
ISBN: 1119132657

A comprehensive look at state-of-the-art ADP theory and real-world applications This book fills a gap in the literature by providing a theoretical framework for integrating techniques from adaptive dynamic programming (ADP) and modern nonlinear control to address data-driven optimal control design challenges arising from both parametric and dynamic uncertainties. Traditional model-based approaches leave much to be desired when addressing the challenges posed by the ever-increasing complexity of real-world engineering systems. An alternative which has received much interest in recent years are biologically-inspired approaches, primarily RADP. Despite their growing popularity worldwide, until now books on ADP have focused nearly exclusively on analysis and design, with scant consideration given to how it can be applied to address robustness issues, a new challenge arising from dynamic uncertainties encountered in common engineering problems. Robust Adaptive Dynamic Programming zeros in on the practical concerns of engineers. The authors develop RADP theory from linear systems to partially-linear, large-scale, and completely nonlinear systems. They provide in-depth coverage of state-of-the-art applications in power systems, supplemented with numerous real-world examples implemented in MATLAB. They also explore fascinating reverse engineering topics, such how ADP theory can be applied to the study of the human brain and cognition. In addition, the book: Covers the latest developments in RADP theory and applications for solving a range of systems’ complexity problems Explores multiple real-world implementations in power systems with illustrative examples backed up by reusable MATLAB code and Simulink block sets Provides an overview of nonlinear control, machine learning, and dynamic control Features discussions of novel applications for RADP theory, including an entire chapter on how it can be used as a computational mechanism of human movement control Robust Adaptive Dynamic Programming is both a valuable working resource and an intriguing exploration of contemporary ADP theory and applications for practicing engineers and advanced students in systems theory, control engineering, computer science, and applied mathematics.

Intelligent Systems

Intelligent Systems
Author: Bogdan M. Wilamowski
Publisher: CRC Press
Total Pages: 610
Release: 2018-10-03
Genre: Technology & Engineering
ISBN: 143980284X

The Industrial Electronics Handbook, Second Edition combines traditional and newer, more specialized knowledge that will help industrial electronics engineers develop practical solutions for the design and implementation of high-power applications. Embracing the broad technological scope of the field, this collection explores fundamental areas, including analog and digital circuits, electronics, electromagnetic machines, signal processing, and industrial control and communications systems. It also facilitates the use of intelligent systems—such as neural networks, fuzzy systems, and evolutionary methods—in terms of a hierarchical structure that makes factory control and supervision more efficient by addressing the needs of all production components. Enhancing its value, this fully updated collection presents research and global trends as published in the IEEE Transactions on Industrial Electronics Journal, one of the largest and most respected publications in the field. As intelligent systems continue to replace and sometimes outperform human intelligence in decision-making processes, they have made substantial contributions to the solution of very complex problems. As a result, the field of computational intelligence has branched out in several directions. For instance, artificial neural networks can learn how to classify patterns, such as images or sequences of events, and effectively model complex nonlinear systems. Simple and easy to implement, fuzzy systems can be applied to successful modeling and system control. Illustrating how these and other tools help engineers model nonlinear system behavior, determine and evaluate system parameters, and ensure overall system control, Intelligent Systems: Addresses various aspects of neural networks and fuzzy systems Focuses on system optimization, covering new techniques such as evolutionary methods, swarm, and ant colony optimizations Discusses several applications that deal with methods of computational intelligence Other volumes in the set: Fundamentals of Industrial Electronics Power Electronics and Motor Drives Control and Mechatronics Industrial Communication Systems

Reinforcement Learning and Approximate Dynamic Programming for Feedback Control

Reinforcement Learning and Approximate Dynamic Programming for Feedback Control
Author: Frank L. Lewis
Publisher: John Wiley & Sons
Total Pages: 498
Release: 2013-01-28
Genre: Technology & Engineering
ISBN: 1118453972

Reinforcement learning (RL) and adaptive dynamic programming (ADP) has been one of the most critical research fields in science and engineering for modern complex systems. This book describes the latest RL and ADP techniques for decision and control in human engineered systems, covering both single player decision and control and multi-player games. Edited by the pioneers of RL and ADP research, the book brings together ideas and methods from many fields and provides an important and timely guidance on controlling a wide variety of systems, such as robots, industrial processes, and economic decision-making.

Optimal Event-Triggered Control Using Adaptive Dynamic Programming

Optimal Event-Triggered Control Using Adaptive Dynamic Programming
Author: Sarangapani Jagannathan
Publisher: CRC Press
Total Pages: 348
Release: 2024-06-21
Genre: Technology & Engineering
ISBN: 1040049168

Optimal Event-triggered Control using Adaptive Dynamic Programming discusses event triggered controller design which includes optimal control and event sampling design for linear and nonlinear dynamic systems including networked control systems (NCS) when the system dynamics are both known and uncertain. The NCS are a first step to realize cyber-physical systems (CPS) or industry 4.0 vision. The authors apply several powerful modern control techniques to the design of event-triggered controllers and derive event-trigger condition and demonstrate closed-loop stability. Detailed derivations, rigorous stability proofs, computer simulation examples, and downloadable MATLAB® codes are included for each case. The book begins by providing background on linear and nonlinear systems, NCS, networked imperfections, distributed systems, adaptive dynamic programming and optimal control, stability theory, and optimal adaptive event-triggered controller design in continuous-time and discrete-time for linear, nonlinear and distributed systems. It lays the foundation for reinforcement learning-based optimal adaptive controller use for infinite horizons. The text then: Introduces event triggered control of linear and nonlinear systems, describing the design of adaptive controllers for them Presents neural network-based optimal adaptive control and game theoretic formulation of linear and nonlinear systems enclosed by a communication network Addresses the stochastic optimal control of linear and nonlinear NCS by using neuro dynamic programming Explores optimal adaptive design for nonlinear two-player zero-sum games under communication constraints to solve optimal policy and event trigger condition Treats an event-sampled distributed linear and nonlinear systems to minimize transmission of state and control signals within the feedback loop via the communication network Covers several examples along the way and provides applications of event triggered control of robot manipulators, UAV and distributed joint optimal network scheduling and control design for wireless NCS/CPS in order to realize industry 4.0 vision An ideal textbook for senior undergraduate students, graduate students, university researchers, and practicing engineers, Optimal Event Triggered Control Design using Adaptive Dynamic Programming instills a solid understanding of neural network-based optimal controllers under event-sampling and how to build them so as to attain CPS or Industry 4.0 vision.

Learning-Based Control

Learning-Based Control
Author: Zhong-Ping Jiang
Publisher: Now Publishers
Total Pages: 122
Release: 2020-12-07
Genre: Technology & Engineering
ISBN: 9781680837520

The recent success of Reinforcement Learning and related methods can be attributed to several key factors. First, it is driven by reward signals obtained through the interaction with the environment. Second, it is closely related to the human learning behavior. Third, it has a solid mathematical foundation. Nonetheless, conventional Reinforcement Learning theory exhibits some shortcomings particularly in a continuous environment or in considering the stability and robustness of the controlled process. In this monograph, the authors build on Reinforcement Learning to present a learning-based approach for controlling dynamical systems from real-time data and review some major developments in this relatively young field. In doing so the authors develop a framework for learning-based control theory that shows how to learn directly suboptimal controllers from input-output data. There are three main challenges on the development of learning-based control. First, there is a need to generalize existing recursive methods. Second, as a fundamental difference between learning-based control and Reinforcement Learning, stability and robustness are important issues that must be addressed for the safety-critical engineering systems such as self-driving cars. Third, data efficiency of Reinforcement Learning algorithms need be addressed for safety-critical engineering systems. This monograph provides the reader with an accessible primer on a new direction in control theory still in its infancy, namely Learning-Based Control Theory, that is closely tied to the literature of safe Reinforcement Learning and Adaptive Dynamic Programming.