Oracle Inequalities in Empirical Risk Minimization and Sparse Recovery Problems

Oracle Inequalities in Empirical Risk Minimization and Sparse Recovery Problems
Author: Vladimir Koltchinskii
Publisher: Springer
Total Pages: 259
Release: 2011-07-29
Genre: Mathematics
ISBN: 3642221475

The purpose of these lecture notes is to provide an introduction to the general theory of empirical risk minimization with an emphasis on excess risk bounds and oracle inequalities in penalized problems. In recent years, there have been new developments in this area motivated by the study of new classes of methods in machine learning such as large margin classification methods (boosting, kernel machines). The main probabilistic tools involved in the analysis of these problems are concentration and deviation inequalities by Talagrand along with other methods of empirical processes theory (symmetrization inequalities, contraction inequality for Rademacher sums, entropy and generic chaining bounds). Sparse recovery based on l_1-type penalization and low rank matrix recovery based on the nuclear norm penalization are other active areas of research, where the main problems can be stated in the framework of penalized empirical risk minimization, and concentration inequalities and empirical processes tools have proved to be very useful.

Oracle Inequalities in Empirical Risk Minimization and Sparse Recovery Problems

Oracle Inequalities in Empirical Risk Minimization and Sparse Recovery Problems
Author: Vladimir Koltchinskii
Publisher: Springer Science & Business Media
Total Pages: 259
Release: 2011-07-29
Genre: Computers
ISBN: 3642221467

The purpose of these lecture notes is to provide an introduction to the general theory of empirical risk minimization with an emphasis on excess risk bounds and oracle inequalities in penalized problems. In recent years, there have been new developments in this area motivated by the study of new classes of methods in machine learning such as large margin classification methods (boosting, kernel machines). The main probabilistic tools involved in the analysis of these problems are concentration and deviation inequalities by Talagrand along with other methods of empirical processes theory (symmetrization inequalities, contraction inequality for Rademacher sums, entropy and generic chaining bounds). Sparse recovery based on l_1-type penalization and low rank matrix recovery based on the nuclear norm penalization are other active areas of research, where the main problems can be stated in the framework of penalized empirical risk minimization, and concentration inequalities and empirical processes tools have proved to be very useful.

Concentration Inequalities

Concentration Inequalities
Author: Stéphane Boucheron
Publisher: Oxford University Press
Total Pages: 492
Release: 2013-02-07
Genre: Mathematics
ISBN: 0199535256

Describes the interplay between the probabilistic structure (independence) and a variety of tools ranging from functional inequalities to transportation arguments to information theory. Applications to the study of empirical processes, random projections, random matrix theory, and threshold phenomena are also presented.

Estimation and Testing Under Sparsity

Estimation and Testing Under Sparsity
Author: Sara van de Geer
Publisher: Springer
Total Pages: 278
Release: 2016-06-28
Genre: Mathematics
ISBN: 3319327747

Taking the Lasso method as its starting point, this book describes the main ingredients needed to study general loss functions and sparsity-inducing regularizers. It also provides a semi-parametric approach to establishing confidence intervals and tests. Sparsity-inducing methods have proven to be very useful in the analysis of high-dimensional data. Examples include the Lasso and group Lasso methods, and the least squares method with other norm-penalties, such as the nuclear norm. The illustrations provided include generalized linear models, density estimation, matrix completion and sparse principal components. Each chapter ends with a problem section. The book can be used as a textbook for a graduate or PhD course.

Empirical Inference

Empirical Inference
Author: Bernhard Schölkopf
Publisher: Springer Science & Business Media
Total Pages: 295
Release: 2013-12-11
Genre: Computers
ISBN: 3642411363

This book honours the outstanding contributions of Vladimir Vapnik, a rare example of a scientist for whom the following statements hold true simultaneously: his work led to the inception of a new field of research, the theory of statistical learning and empirical inference; he has lived to see the field blossom; and he is still as active as ever. He started analyzing learning algorithms in the 1960s and he invented the first version of the generalized portrait algorithm. He later developed one of the most successful methods in machine learning, the support vector machine (SVM) – more than just an algorithm, this was a new approach to learning problems, pioneering the use of functional analysis and convex optimization in machine learning. Part I of this book contains three chapters describing and witnessing some of Vladimir Vapnik's contributions to science. In the first chapter, Léon Bottou discusses the seminal paper published in 1968 by Vapnik and Chervonenkis that lay the foundations of statistical learning theory, and the second chapter is an English-language translation of that original paper. In the third chapter, Alexey Chervonenkis presents a first-hand account of the early history of SVMs and valuable insights into the first steps in the development of the SVM in the framework of the generalised portrait method. The remaining chapters, by leading scientists in domains such as statistics, theoretical computer science, and mathematics, address substantial topics in the theory and practice of statistical learning theory, including SVMs and other kernel-based methods, boosting, PAC-Bayesian theory, online and transductive learning, loss functions, learnable function classes, notions of complexity for function classes, multitask learning, and hypothesis selection. These contributions include historical and context notes, short surveys, and comments on future research directions. This book will be of interest to researchers, engineers, and graduate students engaged with all aspects of statistical learning.

Compressed Sensing and Its Applications

Compressed Sensing and Its Applications
Author: Holger Boche
Publisher: Birkhäuser
Total Pages: 305
Release: 2019-08-13
Genre: Mathematics
ISBN: 3319730746

The chapters in this volume highlight the state-of-the-art of compressed sensing and are based on talks given at the third international MATHEON conference on the same topic, held from December 4-8, 2017 at the Technical University in Berlin. In addition to methods in compressed sensing, chapters provide insights into cutting edge applications of deep learning in data science, highlighting the overlapping ideas and methods that connect the fields of compressed sensing and deep learning. Specific topics covered include: Quantized compressed sensing Classification Machine learning Oracle inequalities Non-convex optimization Image reconstruction Statistical learning theory This volume will be a valuable resource for graduate students and researchers in the areas of mathematics, computer science, and engineering, as well as other applied scientists exploring potential applications of compressed sensing.

Geometric Aspects of Functional Analysis

Geometric Aspects of Functional Analysis
Author: Bo'az Klartag
Publisher: Springer
Total Pages: 366
Release: 2017-04-17
Genre: Mathematics
ISBN: 3319452827

As in the previous Seminar Notes, the current volume reflects general trends in the study of Geometric Aspects of Functional Analysis, understood in a broad sense. A classical theme in the Local Theory of Banach Spaces which is well represented in this volume is the identification of lower-dimensional structures in high-dimensional objects. More recent applications of high-dimensionality are manifested by contributions in Random Matrix Theory, Concentration of Measure and Empirical Processes. Naturally, the Gaussian measure plays a central role in many of these topics, and is also studied in this volume; in particular, the recent breakthrough proof of the Gaussian Correlation Conjecture is revisited. The interplay of the theory with Harmonic and Spectral Analysis is also well apparent in several contributions. The classical relation to both the primal and dual Brunn-Minkowski theories is also well represented, and related algebraic structures pertaining to valuations and valent functions are discussed. All contributions are original research papers and were subject to the usual refereeing standards.

Algorithmic Learning Theory

Algorithmic Learning Theory
Author: Ronald Ortner
Publisher: Springer
Total Pages: 382
Release: 2016-10-12
Genre: Computers
ISBN: 3319463799

This book constitutes the refereed proceedings of the 27th International Conference on Algorithmic Learning Theory, ALT 2016, held in Bari, Italy, in October 2016, co-located with the 19th International Conference on Discovery Science, DS 2016. The 24 regular papers presented in this volume were carefully reviewed and selected from 45 submissions. In addition the book contains 5 abstracts of invited talks. The papers are organized in topical sections named: error bounds, sample compression schemes; statistical learning, theory, evolvability; exact and interactive learning; complexity of teaching models; inductive inference; online learning; bandits and reinforcement learning; and clustering.

High Dimensional Probability VI

High Dimensional Probability VI
Author: Christian Houdré
Publisher: Springer Science & Business Media
Total Pages: 372
Release: 2013-04-19
Genre: Mathematics
ISBN: 3034804903

This is a collection of papers by participants at High Dimensional Probability VI Meeting held from October 9-14, 2011 at the Banff International Research Station in Banff, Alberta, Canada. High Dimensional Probability (HDP) is an area of mathematics that includes the study of probability distributions and limit theorems in infinite-dimensional spaces such as Hilbert spaces and Banach spaces. The most remarkable feature of this area is that it has resulted in the creation of powerful new tools and perspectives, whose range of application has led to interactions with other areas of mathematics, statistics, and computer science. These include random matrix theory, nonparametric statistics, empirical process theory, statistical learning theory, concentration of measure phenomena, strong and weak approximations, distribution function estimation in high dimensions, combinatorial optimization, and random graph theory. The papers in this volume show that HDP theory continues to develop new tools, methods, techniques and perspectives to analyze the random phenomena. Both researchers and advanced students will find this book of great use for learning about new avenues of research.​