Continuous-time Stochastic Control and Optimization with Financial Applications

Continuous-time Stochastic Control and Optimization with Financial Applications
Author: Huyên Pham
Publisher: Springer Science & Business Media
Total Pages: 243
Release: 2009-05-28
Genre: Mathematics
ISBN: 3540895000

Stochastic optimization problems arise in decision-making problems under uncertainty, and find various applications in economics and finance. On the other hand, problems in finance have recently led to new developments in the theory of stochastic control. This volume provides a systematic treatment of stochastic optimization problems applied to finance by presenting the different existing methods: dynamic programming, viscosity solutions, backward stochastic differential equations, and martingale duality methods. The theory is discussed in the context of recent developments in this field, with complete and detailed proofs, and is illustrated by means of concrete examples from the world of finance: portfolio allocation, option hedging, real options, optimal investment, etc. This book is directed towards graduate students and researchers in mathematical finance, and will also benefit applied mathematicians interested in financial applications and practitioners wishing to know more about the use of stochastic optimization methods in finance.

Stochastic Controls

Stochastic Controls
Author: Jiongmin Yong
Publisher: Springer Science & Business Media
Total Pages: 459
Release: 2012-12-06
Genre: Mathematics
ISBN: 1461214661

As is well known, Pontryagin's maximum principle and Bellman's dynamic programming are the two principal and most commonly used approaches in solving stochastic optimal control problems. * An interesting phenomenon one can observe from the literature is that these two approaches have been developed separately and independently. Since both methods are used to investigate the same problems, a natural question one will ask is the fol lowing: (Q) What is the relationship betwccn the maximum principlc and dy namic programming in stochastic optimal controls? There did exist some researches (prior to the 1980s) on the relationship between these two. Nevertheless, the results usually werestated in heuristic terms and proved under rather restrictive assumptions, which were not satisfied in most cases. In the statement of a Pontryagin-type maximum principle there is an adjoint equation, which is an ordinary differential equation (ODE) in the (finite-dimensional) deterministic case and a stochastic differential equation (SDE) in the stochastic case. The system consisting of the adjoint equa tion, the original state equation, and the maximum condition is referred to as an (extended) Hamiltonian system. On the other hand, in Bellman's dynamic programming, there is a partial differential equation (PDE), of first order in the (finite-dimensional) deterministic case and of second or der in the stochastic case. This is known as a Hamilton-Jacobi-Bellman (HJB) equation.

Stochastic Optimization in Continuous Time

Stochastic Optimization in Continuous Time
Author: Fwu-Ranq Chang
Publisher: Cambridge University Press
Total Pages: 346
Release: 2004-04-26
Genre: Business & Economics
ISBN: 1139452223

First published in 2004, this is a rigorous but user-friendly book on the application of stochastic control theory to economics. A distinctive feature of the book is that mathematical concepts are introduced in a language and terminology familiar to graduate students of economics. The standard topics of many mathematics, economics and finance books are illustrated with real examples documented in the economic literature. Moreover, the book emphasises the dos and don'ts of stochastic calculus, cautioning the reader that certain results and intuitions cherished by many economists do not extend to stochastic models. A special chapter (Chapter 5) is devoted to exploring various methods of finding a closed-form representation of the value function of a stochastic control problem, which is essential for ascertaining the optimal policy functions. The book also includes many practice exercises for the reader. Notes and suggested readings are provided at the end of each chapter for more references and possible extensions.

Optimization, Control, and Applications of Stochastic Systems

Optimization, Control, and Applications of Stochastic Systems
Author: Daniel Hernández-Hernández
Publisher: Springer Science & Business Media
Total Pages: 331
Release: 2012-08-15
Genre: Science
ISBN: 0817683372

This volume provides a general overview of discrete- and continuous-time Markov control processes and stochastic games, along with a look at the range of applications of stochastic control and some of its recent theoretical developments. These topics include various aspects of dynamic programming, approximation algorithms, and infinite-dimensional linear programming. In all, the work comprises 18 carefully selected papers written by experts in their respective fields. Optimization, Control, and Applications of Stochastic Systems will be a valuable resource for all practitioners, researchers, and professionals in applied mathematics and operations research who work in the areas of stochastic control, mathematical finance, queueing theory, and inventory systems. It may also serve as a supplemental text for graduate courses in optimal control and dynamic games.

Controlled Markov Processes and Viscosity Solutions

Controlled Markov Processes and Viscosity Solutions
Author: Wendell H. Fleming
Publisher: Springer Science & Business Media
Total Pages: 436
Release: 2006-02-04
Genre: Mathematics
ISBN: 0387310711

This book is an introduction to optimal stochastic control for continuous time Markov processes and the theory of viscosity solutions. It covers dynamic programming for deterministic optimal control problems, as well as to the corresponding theory of viscosity solutions. New chapters in this second edition introduce the role of stochastic optimal control in portfolio optimization and in pricing derivatives in incomplete markets and two-controller, zero-sum differential games.

Optimization, Control, and Applications of Stochastic Systems

Optimization, Control, and Applications of Stochastic Systems
Author: Daniel Hernández-Hernández
Publisher: Birkhäuser
Total Pages: 309
Release: 2012-08-14
Genre: Science
ISBN: 9780817683368

This volume provides a general overview of discrete- and continuous-time Markov control processes and stochastic games, along with a look at the range of applications of stochastic control and some of its recent theoretical developments. These topics include various aspects of dynamic programming, approximation algorithms, and infinite-dimensional linear programming. In all, the work comprises 18 carefully selected papers written by experts in their respective fields. Optimization, Control, and Applications of Stochastic Systems will be a valuable resource for all practitioners, researchers, and professionals in applied mathematics and operations research who work in the areas of stochastic control, mathematical finance, queueing theory, and inventory systems. It may also serve as a supplemental text for graduate courses in optimal control and dynamic games.

Stochastic Control in Insurance

Stochastic Control in Insurance
Author: Hanspeter Schmidli
Publisher: Springer Science & Business Media
Total Pages: 263
Release: 2007-11-20
Genre: Business & Economics
ISBN: 1848000030

Yet again, here is a Springer volume that offers readers something completely new. Until now, solved examples of the application of stochastic control to actuarial problems could only be found in journals. Not any more: this is the first book to systematically present these methods in one volume. The author starts with a short introduction to stochastic control techniques, then applies the principles to several problems. These examples show how verification theorems and existence theorems may be proved, and that the non-diffusion case is simpler than the diffusion case. Schmidli’s brilliant text also includes a number of appendices, a vital resource for those in both academic and professional settings.

Relative Optimization of Continuous-Time and Continuous-State Stochastic Systems

Relative Optimization of Continuous-Time and Continuous-State Stochastic Systems
Author: Xi-Ren Cao
Publisher: Springer Nature
Total Pages: 376
Release: 2020-05-13
Genre: Technology & Engineering
ISBN: 3030418464

This monograph applies the relative optimization approach to time nonhomogeneous continuous-time and continuous-state dynamic systems. The approach is intuitively clear and does not require deep knowledge of the mathematics of partial differential equations. The topics covered have the following distinguishing features: long-run average with no under-selectivity, non-smooth value functions with no viscosity solutions, diffusion processes with degenerate points, multi-class optimization with state classification, and optimization with no dynamic programming. The book begins with an introduction to relative optimization, including a comparison with the traditional approach of dynamic programming. The text then studies the Markov process, focusing on infinite-horizon optimization problems, and moves on to discuss optimal control of diffusion processes with semi-smooth value functions and degenerate points, and optimization of multi-dimensional diffusion processes. The book concludes with a brief overview of performance derivative-based optimization. Among the more important novel considerations presented are: the extension of the Hamilton–Jacobi–Bellman optimality condition from smooth to semi-smooth value functions by derivation of explicit optimality conditions at semi-smooth points and application of this result to degenerate and reflected processes; proof of semi-smoothness of the value function at degenerate points; attention to the under-selectivity issue for the long-run average and bias optimality; discussion of state classification for time nonhomogeneous continuous processes and multi-class optimization; and development of the multi-dimensional Tanaka formula for semi-smooth functions and application of this formula to stochastic control of multi-dimensional systems with degenerate points. The book will be of interest to researchers and students in the field of stochastic control and performance optimization alike.

Optimal Stochastic Control, Stochastic Target Problems, and Backward SDE

Optimal Stochastic Control, Stochastic Target Problems, and Backward SDE
Author: Nizar Touzi
Publisher: Springer Science & Business Media
Total Pages: 219
Release: 2012-09-27
Genre: Mathematics
ISBN: 1461442850

​This book collects some recent developments in stochastic control theory with applications to financial mathematics. We first address standard stochastic control problems from the viewpoint of the recently developed weak dynamic programming principle. A special emphasis is put on the regularity issues and, in particular, on the behavior of the value function near the boundary. We then provide a quick review of the main tools from viscosity solutions which allow to overcome all regularity problems. We next address the class of stochastic target problems which extends in a nontrivial way the standard stochastic control problems. Here the theory of viscosity solutions plays a crucial role in the derivation of the dynamic programming equation as the infinitesimal counterpart of the corresponding geometric dynamic programming equation. The various developments of this theory have been stimulated by applications in finance and by relevant connections with geometric flows. Namely, the second order extension was motivated by illiquidity modeling, and the controlled loss version was introduced following the problem of quantile hedging. The third part specializes to an overview of Backward stochastic differential equations, and their extensions to the quadratic case.​

Numerical Methods for Stochastic Control Problems in Continuous Time

Numerical Methods for Stochastic Control Problems in Continuous Time
Author: Harold Kushner
Publisher: Springer Science & Business Media
Total Pages: 480
Release: 2013-11-27
Genre: Mathematics
ISBN: 146130007X

Stochastic control is a very active area of research. This monograph, written by two leading authorities in the field, has been updated to reflect the latest developments. It covers effective numerical methods for stochastic control problems in continuous time on two levels, that of practice and that of mathematical development. It is broadly accessible for graduate students and researchers.