Online Model-Free Distributed Reinforcement Learning Approach for Networked Systems of Self-organizing Agents

Online Model-Free Distributed Reinforcement Learning Approach for Networked Systems of Self-organizing Agents
Author: Yiqing Chen
Publisher:
Total Pages:
Release: 2021
Genre:
ISBN:

Control of large groups of robotic agents is driven by applications including military, aeronautics and astronautics, transportation network, and environmental monitoring. Cooperative control of networked multi-agent systems aims at driving the behavior of the group via feedback control inputs that encode the groups' dynamics based on information sharing, with inter-agent communications that can be time varying and be spatially non-uniform. Notably, local interaction rules can induce coordinated behaviour, provided suitable network topologies. Distributed learning paradigms are often necessary for this class of systems to be able to operate autonomously and robustly, without the need of external units providing centralized information. Compared with model-based protocols that can be computationally prohibitive due to their mathematical complexity and requirements in terms of feedback information, we present an online model-free algorithm for some nonlinear tracking problems with unknown system dynamics. This method prescribes the actuation forces of agents to follow the time-varying trajectory of a moving target. The tracking problem is addressed by an online value iteration process which requires measurements collected along the trajectories. A set of simulations are conducted to illustrate that the presented algorithm is well functioning in various reference-tracking scenarios.

Distributed Economic Operation in Smart Grid: Model-Based and Model-Free Perspectives

Distributed Economic Operation in Smart Grid: Model-Based and Model-Free Perspectives
Author: Jiahu Qin
Publisher: Springer Nature
Total Pages: 246
Release: 2023-01-25
Genre: Technology & Engineering
ISBN: 9811985944

This book aims to work out the distributed economic operation in smart grids in a systematic way, which ranges from model-based to model-free perspectives. The main contributions of this book can be summarized into three folds. First, we investigate the fundamental economic operation problems in smart grids from model-based perspective. Specifically, these problems can be modeled as deterministic optimization models, and we propose some distributed optimization algorithms by integrating the multi-agent consensus theory and optimization techniques to achieve the distributed coordination of various generation units and loads. Second, due to the randomness of the large-scale renewable energies and the flexibility of the loads, we further address these economic operation problems from a model-free perspective, and we propose learning-based approaches to address the uncertainty and randomness. At last, we extend the idea of model-based and model-free algorithms to plug-in electric vehicles (PEVs) charging/discharging scheduling problem, the key challenge of which involves multiple objectives simultaneously while the behavior of PEVs and the electricity price are intrinsically random. This book presents several recent theoretical findings on distributed economic operation in smart grids from model-based and model-free perspectives. By systematically integrating novel ideas, fresh insights, and rigorous results, this book provides a base for further theoretical research on distributed economic operation in smart grids. It can be a reference for graduates and researchers to study the operation and management in smart grids. Some prerequisites for reading this book include optimization theory, matrix theory, game theory, reinforcement learning, etc.

Distributed Algorithms for Self-adapting and Self-organizing Systems

Distributed Algorithms for Self-adapting and Self-organizing Systems
Author: John C. Meyer
Publisher:
Total Pages: 356
Release: 2013
Genre: Adaptive control systems
ISBN:

Over the last half century computing power and the proliferation of computing devices have increased at an exponential rate. Smart phones are becoming more prevalent allowing access to other people, other systems, and information. This growth coupled with the explosion of networking and the Internet has tied these disparate devices together. At the present rate of growth the complexity appears to be approaching the limits of human capability. This increase in complexity has lead to IBM's Grand Challenge in Autonomic Computing. In autonomic computing, computers should be able to handle mundane tasks themselves, adapting to their environment, allowing users to specify their needs at a higher level. Two areas of autonomic computing dominate current research in industry and academia in response to the challenge, namely: Data centers, and distributed computing. Much progress has been made related to specific problems in these domains including power management, visualization, work load distribution, and scalability. However, much remains to be done in the areas of design, trustworthiness, validation, and certification of these systems. This dissertation aims to contribute to addressing the challenge. It proposes a method to specify the needs of a system and then translate those specifications into an algorithm for a distributed, networked, federated system that meets the specified goals. To validate the approach case studies in both self-adapting and self-organizing systems are explored. The first is a self-adaptive system which manages a scarce resource in a changing environment. As the environment changes the system adapts to optimize resource usage. The next two use agents who collaborate to self-organize to perform the function desired in an adaptive fashion. The design inspiration in these approaches comes from two different natural paradigms: Newtonian forces and potential energy. The case studies are assessed and validated through a series of qualitative and quantitative metrics.

Design and Control of Self-organizing Systems

Design and Control of Self-organizing Systems
Author: Carlos Gershenson
Publisher: CopIt ArXives
Total Pages: 189
Release: 2007-09-05
Genre: Science
ISBN: 0983117233

Complex systems are usually difficult to design and control. There are several particular methods for coping with complexity, but there is no general approach to build complex systems. In this book I propose a methodology to aid engineers in the design and control of complex systems. This is based on the description of systems as self-organizing. Starting from the agent metaphor, the methodology proposes a conceptual framework and a series of steps to follow to find proper mechanisms that will promote elements to find solutions by actively interacting among themselves.

Reinforcement Learning, second edition

Reinforcement Learning, second edition
Author: Richard S. Sutton
Publisher: MIT Press
Total Pages: 549
Release: 2018-11-13
Genre: Computers
ISBN: 0262352702

The significantly expanded and updated new edition of a widely used text on reinforcement learning, one of the most active research areas in artificial intelligence. Reinforcement learning, one of the most active research areas in artificial intelligence, is a computational approach to learning whereby an agent tries to maximize the total amount of reward it receives while interacting with a complex, uncertain environment. In Reinforcement Learning, Richard Sutton and Andrew Barto provide a clear and simple account of the field's key ideas and algorithms. This second edition has been significantly expanded and updated, presenting new topics and updating coverage of other topics. Like the first edition, this second edition focuses on core online learning algorithms, with the more mathematical material set off in shaded boxes. Part I covers as much of reinforcement learning as possible without going beyond the tabular case for which exact solutions can be found. Many algorithms presented in this part are new to the second edition, including UCB, Expected Sarsa, and Double Learning. Part II extends these ideas to function approximation, with new sections on such topics as artificial neural networks and the Fourier basis, and offers expanded treatment of off-policy learning and policy-gradient methods. Part III has new chapters on reinforcement learning's relationships to psychology and neuroscience, as well as an updated case-studies chapter including AlphaGo and AlphaGo Zero, Atari game playing, and IBM Watson's wagering strategy. The final chapter discusses the future societal impacts of reinforcement learning.

Advances in Practical Applications of Cyber-Physical Multi-Agent Systems: The PAAMS Collection

Advances in Practical Applications of Cyber-Physical Multi-Agent Systems: The PAAMS Collection
Author: Yves Demazeau
Publisher: Springer
Total Pages: 384
Release: 2017-06-08
Genre: Computers
ISBN: 3319599305

This book constitutes the refereed proceedings of the 15th International Conference on Practical Applications of Scalable Multi-Agent Systems, PAAMS 2017, held in Porto, Portugal, in June 2017. The 11 revised full papers, 11 short papers, and 17 Demo papers were carefully reviewed and selected from 63 submissions. The papers report on the application and validation of agent-based models, methods, and technologies in a number of key application areas, including day life and real world, energy and networks, human and trust, markets and bids, models and tools, negotiation and conversation, scalability and resources.

Advanced Technologies for Planning and Operation of Prosumer Energy Systems, volume III

Advanced Technologies for Planning and Operation of Prosumer Energy Systems, volume III
Author: Bin Zhou
Publisher: Frontiers Media SA
Total Pages: 385
Release: 2024-07-30
Genre: Technology & Engineering
ISBN: 2832552463

Prosumers, such as energy storage, smart home, and microgrids, are the consumers who also produce and share surplus energy with other users. With capabilities of flexibly managing the generation, storage and consumption of energy in a simultaneous manner, prosumers can help improve the operation efficiency of smart grid. Due to the rapid expansion of prosumer clusters, the planning and operation issues of prosumer energy systems have been increasingly raised. Aspects including energy infrastructure design, energy management, system stability, etc., are urgently required to be addressed while taking full advantage of prosumers' capabilities. However, up to date, the research on prosumers has not drawn sufficient attention. This proposal presents the need to introduce a Research Topic on prosumer energy systems in Frontiers in Energy Research. We believe this Research Topic can promote the research on advanced planning and operation technologies of prosumer energy systems and contribute to the carbon neutrality for a sustainable society.

Scaling Multi-agent Learning in Complex Environments

Scaling Multi-agent Learning in Complex Environments
Author: Chongjie Zhang
Publisher:
Total Pages: 194
Release: 2011
Genre: Computer-assisted instruction
ISBN:

Cooperative multi-agent systems (MAS) are finding applications in a wide variety of domains, including sensor networks, robotics, distributed control, collaborative decision support systems, and data mining. A cooperative MAS consists of a group of autonomous agents that interact with one another in order to optimize a global performance measure. A central challenge in cooperative MAS research is to design distributed coordination policies. Designing optimal distributed coordination policies offline is usually not feasible for large-scale complex multi-agent systems, where 10s to 1000s of agents are involved, there is limited communication bandwidth and communication delay between agents, agents have only limited partial views of the whole system, etc. This infeasibility is either due to a prohibitive cost to build an accurate decision model, or a dynamically evolving environment, or the intractable computation complexity. This thesis develops a multi-agent reinforcement learning paradigm to allow agents to effectively learn and adapt coordination policies in complex cooperative domains without explicitly building the complete decision models. With multi-agent reinforcement learning (MARL), agents explore the environment through trial and error, adapt their behaviors to the dynamics of the uncertain and evolving environment, and improve their performance through experiences. To achieve the scalability of MARL and ensure the global performance, the MARL paradigm developed in this thesis restricts the learning of each agent to using information locally observed or received from local interactions with a limited number of agents (i.e., neighbors) in the system and exploits non-local interaction information to coordinate the learning processes of agents. This thesis develops new MARL algorithms for agents to learn effectively with limited observations in multi-agent settings and introduces a low-overhead supervisory control framework to collect and integrate non-local information into the learning process of agents to coordinate their learning. More specifically, the contributions of already completed aspects of this thesis are as follows: Multi-Agent Learning with Policy Prediction: This thesis introduces the concept of policy prediction and augments the basic gradient-based learning algorithm to achieve two properties: best-response learning and convergence. The convergence property of multi-agent learning with policy prediction is proven for a class of static games under the assumption of full observability. MARL Algorithm with Limited Observability: This thesis develops PGA-APP, a practical multi-agent learning algorithm that extends Q-learning to learn stochastic policies. PGA-APP combines the policy gradient technique with the idea of policy prediction. It allows an agent to learn effectively with limited observability in complex domains in presence of other learning agents. The empirical results demonstrate that PGA-APP outperforms state-of-the-art MARL techniques in both benchmark games. MARL Application in Cloud Computing: This thesis illustrates how MARL can be applied to optimizing online distributed resource allocation in cloud computing. Empirical results show that the MARL approach performs reasonably well, compared to an optimal solution, and better than a centralized myopic allocation approach in some cases. A General Paradigm for Coordinating MARL: This thesis presents a multi-level supervisory control framework to coordinate and guide the agents' learning process. This framework exploits non-local information and introduces a more global view to coordinate the learning process of individual agents without incurring significant overhead and exploding their policy space. Empirical results demonstrate that this coordination significantly improves the speed, quality and likelihood of MARL convergence in large-scale, complex cooperative multi-agent systems. An Agent Interaction Model: This thesis proposes a new general agent interaction model. This interaction model formalizes a type of interactions among agents, called {\em joint-even-driven} interactions, and define a measure for capturing the strength of such interactions. Formal analysis reveals the relationship between interactions between agents and the performance of individual agents and the whole system. Self-Organization for Nearly-Decomposable Hierarchy: This thesis develops a distributed self-organization approach, based on the agent interaction model, that dynamically form a nearly decomposable hierarchy for large-scale multi-agent systems. This self-organization approach is integrated into supervisory control framework to automatically evolving supervisory organizations to better coordinating MARL during the learning process. Empirically results show that dynamically evolving supervisory organizations can perform better than static ones. Automating Coordination for Multi-Agent Learning: We tailor our supervision framework for coordinating MARL in ND-POMDPs. By exploiting structured interaction in ND-POMDPs, this tailored approach distributes the learning of the global joint policy among supervisors and employs DCOP techniques to automatically coordinate distributed learning to ensure the global learning performance. We prove that this approach can learn a globally optimal policy for ND-POMDPs with a property called groupwise observability.

Process Planning and Scheduling for Distributed Manufacturing

Process Planning and Scheduling for Distributed Manufacturing
Author: Lihui Wang
Publisher: Springer Science & Business Media
Total Pages: 441
Release: 2007-05-14
Genre: Technology & Engineering
ISBN: 1846287529

This is the first book to focus on emerging technologies for distributed intelligent decision-making in process planning and dynamic scheduling. It has two sections: a review of several key areas of research, and an in-depth treatment of particular techniques. Each chapter addresses a specific problem domain and offers practical solutions to solve it. The book provides a better understanding of the present state and future trends of research in this area.