Reinforcement Learning and Stochastic Optimization

Reinforcement Learning and Stochastic Optimization
Author: Warren B. Powell
Publisher: John Wiley & Sons
Total Pages: 1090
Release: 2022-03-15
Genre: Mathematics
ISBN: 1119815037

REINFORCEMENT LEARNING AND STOCHASTIC OPTIMIZATION Clearing the jungle of stochastic optimization Sequential decision problems, which consist of “decision, information, decision, information,” are ubiquitous, spanning virtually every human activity ranging from business applications, health (personal and public health, and medical decision making), energy, the sciences, all fields of engineering, finance, and e-commerce. The diversity of applications attracted the attention of at least 15 distinct fields of research, using eight distinct notational systems which produced a vast array of analytical tools. A byproduct is that powerful tools developed in one community may be unknown to other communities. Reinforcement Learning and Stochastic Optimization offers a single canonical framework that can model any sequential decision problem using five core components: state variables, decision variables, exogenous information variables, transition function, and objective function. This book highlights twelve types of uncertainty that might enter any model and pulls together the diverse set of methods for making decisions, known as policies, into four fundamental classes that span every method suggested in the academic literature or used in practice. Reinforcement Learning and Stochastic Optimization is the first book to provide a balanced treatment of the different methods for modeling and solving sequential decision problems, following the style used by most books on machine learning, optimization, and simulation. The presentation is designed for readers with a course in probability and statistics, and an interest in modeling and applications. Linear programming is occasionally used for specific problem classes. The book is designed for readers who are new to the field, as well as those with some background in optimization under uncertainty. Throughout this book, readers will find references to over 100 different applications, spanning pure learning problems, dynamic resource allocation problems, general state-dependent problems, and hybrid learning/resource allocation problems such as those that arose in the COVID pandemic. There are 370 exercises, organized into seven groups, ranging from review questions, modeling, computation, problem solving, theory, programming exercises and a “diary problem” that a reader chooses at the beginning of the book, and which is used as a basis for questions throughout the rest of the book.

Structural Health Monitoring 2013: A Roadmap to Intelligent Structures

Structural Health Monitoring 2013: A Roadmap to Intelligent Structures
Author: Fu-Kuo Chang
Publisher: DEStech Publications, Inc
Total Pages: 1434
Release: 2013-09-26
Genre: Technology & Engineering
ISBN: 1605951153

Original research on SHM sensors, quantification strategies, system integration and control for a wide range of engineered materials New applications in robotics, machinery, as well as military aircraft, railroads, highways, bridges, pipelines, stadiums, tunnels, space exploration and energy production Continuing a critical book series on structural health monitoring (SHM), this two-volume set (with full-text searchable CD-ROM) offers, as its subtitle implies, a guide to greater integration and control of SHM systems. Specifically, the volumes contain new research that will enable readers to more efficiently link sensor detection, diagnostics/quantification, overall system functionality, and automated, e.g., robotic, control, thus further closing the loop from inherent signal-based damage detection to responsive real-time maintenance and repair. SHM performance is demonstrated in monitoring the behavior of composites, metals, concrete, polymers and selected nanomaterials in a wide array of surroundings, including harsh environments, under extreme (e.g., seismic) loading and in space. New information on smart sensors and network optimization is enhanced by novel statistical and model-based methods for signal processing and data quantification. A special feature of the book is its explanation of emerging control technologies. Research in these volumes was initially presented in September 2013 at the 9th International Workshop on Structural Health Monitoring (IWSHM), held at Stanford University and sponsored by the Air Force Office of Scientific Research, the Army Research Laboratory, and the Office of Naval Research.

Bridging the Gap Between AI and Reality

Bridging the Gap Between AI and Reality
Author: Bernhard Steffen
Publisher: Springer Nature
Total Pages: 454
Release: 2023-12-13
Genre: Computers
ISBN: 3031460022

This book constitutes the proceedings of the First International Conference on Bridging the Gap between AI and Reality, AISoLA 2023, which took place in Crete, Greece, in October 2023. The papers included in this book focus on the following topics: The nature of AI-based systems; ethical, economic and legal implications of AI-systems in practice; ways to make controlled use of AI via the various kinds of formal methods-based validation techniques; dedicated applications scenarios which may allow certain levels of assistance; and education in times of deep learning.

Partially Observed Markov Decision Processes

Partially Observed Markov Decision Processes
Author: Vikram Krishnamurthy
Publisher: Cambridge University Press
Total Pages: 491
Release: 2016-03-21
Genre: Mathematics
ISBN: 1107134609

This book covers formulation, algorithms, and structural results of partially observed Markov decision processes, whilst linking theory to real-world applications in controlled sensing. Computations are kept to a minimum, enabling students and researchers in engineering, operations research, and economics to understand the methods and determine the structure of their optimal solution.

Reinforcement Learning, second edition

Reinforcement Learning, second edition
Author: Richard S. Sutton
Publisher: MIT Press
Total Pages: 549
Release: 2018-11-13
Genre: Computers
ISBN: 0262352702

The significantly expanded and updated new edition of a widely used text on reinforcement learning, one of the most active research areas in artificial intelligence. Reinforcement learning, one of the most active research areas in artificial intelligence, is a computational approach to learning whereby an agent tries to maximize the total amount of reward it receives while interacting with a complex, uncertain environment. In Reinforcement Learning, Richard Sutton and Andrew Barto provide a clear and simple account of the field's key ideas and algorithms. This second edition has been significantly expanded and updated, presenting new topics and updating coverage of other topics. Like the first edition, this second edition focuses on core online learning algorithms, with the more mathematical material set off in shaded boxes. Part I covers as much of reinforcement learning as possible without going beyond the tabular case for which exact solutions can be found. Many algorithms presented in this part are new to the second edition, including UCB, Expected Sarsa, and Double Learning. Part II extends these ideas to function approximation, with new sections on such topics as artificial neural networks and the Fourier basis, and offers expanded treatment of off-policy learning and policy-gradient methods. Part III has new chapters on reinforcement learning's relationships to psychology and neuroscience, as well as an updated case-studies chapter including AlphaGo and AlphaGo Zero, Atari game playing, and IBM Watson's wagering strategy. The final chapter discusses the future societal impacts of reinforcement learning.

Multistage Stochastic Optimization

Multistage Stochastic Optimization
Author: Georg Ch. Pflug
Publisher: Springer
Total Pages: 309
Release: 2014-11-12
Genre: Business & Economics
ISBN: 3319088432

Multistage stochastic optimization problems appear in many ways in finance, insurance, energy production and trading, logistics and transportation, among other areas. They describe decision situations under uncertainty and with a longer planning horizon. This book contains a comprehensive treatment of today’s state of the art in multistage stochastic optimization. It covers the mathematical backgrounds of approximation theory as well as numerous practical algorithms and examples for the generation and handling of scenario trees. A special emphasis is put on estimation and bounding of the modeling error using novel distance concepts, on time consistency and the role of model ambiguity in the decision process. An extensive treatment of examples from electricity production, asset liability management and inventory control concludes the book.

A Concise Introduction to Decentralized POMDPs

A Concise Introduction to Decentralized POMDPs
Author: Frans A. Oliehoek
Publisher: Springer
Total Pages: 146
Release: 2016-06-03
Genre: Computers
ISBN: 3319289292

This book introduces multiagent planning under uncertainty as formalized by decentralized partially observable Markov decision processes (Dec-POMDPs). The intended audience is researchers and graduate students working in the fields of artificial intelligence related to sequential decision making: reinforcement learning, decision-theoretic planning for single agents, classical multiagent planning, decentralized control, and operations research.

Algorithms for Reinforcement Learning

Algorithms for Reinforcement Learning
Author: Csaba Grossi
Publisher: Springer Nature
Total Pages: 89
Release: 2022-05-31
Genre: Computers
ISBN: 3031015517

Reinforcement learning is a learning paradigm concerned with learning to control a system so as to maximize a numerical performance measure that expresses a long-term objective. What distinguishes reinforcement learning from supervised learning is that only partial feedback is given to the learner about the learner's predictions. Further, the predictions may have long term effects through influencing the future state of the controlled system. Thus, time plays a special role. The goal in reinforcement learning is to develop efficient learning algorithms, as well as to understand the algorithms' merits and limitations. Reinforcement learning is of great interest because of the large number of practical applications that it can be used to address, ranging from problems in artificial intelligence to operations research or control engineering. In this book, we focus on those algorithms of reinforcement learning that build on the powerful theory of dynamic programming. We give a fairly comprehensive catalog of learning problems, describe the core ideas, note a large number of state of the art algorithms, followed by the discussion of their theoretical properties and limitations. Table of Contents: Markov Decision Processes / Value Prediction Problems / Control / For Further Exploration

Automated Machine Learning

Automated Machine Learning
Author: Frank Hutter
Publisher: Springer
Total Pages: 223
Release: 2019-05-17
Genre: Computers
ISBN: 3030053180

This open access book presents the first comprehensive overview of general methods in Automated Machine Learning (AutoML), collects descriptions of existing systems based on these methods, and discusses the first series of international challenges of AutoML systems. The recent success of commercial ML applications and the rapid growth of the field has created a high demand for off-the-shelf ML methods that can be used easily and without expert knowledge. However, many of the recent machine learning successes crucially rely on human experts, who manually select appropriate ML architectures (deep learning architectures or more traditional ML workflows) and their hyperparameters. To overcome this problem, the field of AutoML targets a progressive automation of machine learning, based on principles from optimization and machine learning itself. This book serves as a point of entry into this quickly-developing field for researchers and advanced students alike, as well as providing a reference for practitioners aiming to use AutoML in their work.