[PDF] Convex Optimization With Computational Errors - eBooks Review

Convex Optimization With Computational Errors


Convex Optimization With Computational Errors
DOWNLOAD

Download Convex Optimization With Computational Errors PDF/ePub or read online books in Mobi eBooks. Click Download or Read Online button to get Convex Optimization With Computational Errors book now. This website allows unlimited access to, at the time of writing, more than 1.5 million titles, including hundreds of thousands of titles in various foreign languages. If the content not found or just blank you must refresh this page



Convex Optimization With Computational Errors


Convex Optimization With Computational Errors
DOWNLOAD
Author : Alexander J. Zaslavski
language : en
Publisher: Springer Nature
Release Date : 2020-01-31

Convex Optimization With Computational Errors written by Alexander J. Zaslavski and has been published by Springer Nature this book supported file pdf, txt, epub, kindle and other format this book has been release on 2020-01-31 with Mathematics categories.


The book is devoted to the study of approximate solutions of optimization problems in the presence of computational errors. It contains a number of results on the convergence behavior of algorithms in a Hilbert space, which are known as important tools for solving optimization problems. The research presented in the book is the continuation and the further development of the author's (c) 2016 book Numerical Optimization with Computational Errors, Springer 2016. Both books study the algorithms taking into account computational errors which are always present in practice. The main goal is, for a known computational error, to find out what an approximate solution can be obtained and how many iterates one needs for this. The main difference between this new book and the 2016 book is that in this present book the discussion takes into consideration the fact that for every algorithm, its iteration consists of several steps and that computational errors for different steps are generally, different. This fact, which was not taken into account in the previous book, is indeed important in practice. For example, the subgradient projection algorithm consists of two steps. The first step is a calculation of a subgradient of the objective function while in the second one we calculate a projection on the feasible set. In each of these two steps there is a computational error and these two computational errors are different in general. It may happen that the feasible set is simple and the objective function is complicated. As a result, the computational error, made when one calculates the projection, is essentially smaller than the computational error of the calculation of the subgradient. Clearly, an opposite case is possible too. Another feature of this book is a study of a number of important algorithms which appeared recently in the literature and which are not discussed in the previous book. This monograph contains 12 chapters. Chapter 1 is an introduction. In Chapter 2 we study the subgradient projection algorithm for minimization of convex and nonsmooth functions. We generalize the results of [NOCE] and establish results which has no prototype in [NOCE]. In Chapter 3 we analyze the mirror descent algorithm for minimization of convex and nonsmooth functions, under the presence of computational errors. For this algorithm each iteration consists of two steps. The first step is a calculation of a subgradient of the objective function while in the second one we solve an auxiliary minimization problem on the set of feasible points. In each of these two steps there is a computational error. We generalize the results of [NOCE] and establish results which has no prototype in [NOCE]. In Chapter 4 we analyze the projected gradient algorithm with a smooth objective function under the presence of computational errors. In Chapter 5 we consider an algorithm, which is an extension of the projection gradient algorithm used for solving linear inverse problems arising in signal/image processing. In Chapter 6 we study continuous subgradient method and continuous subgradient projection algorithm for minimization of convex nonsmooth functions and for computing the saddle points of convex-concave functions, under the presence of computational errors. All the results of this chapter has no prototype in [NOCE]. In Chapters 7-12 we analyze several algorithms under the presence of computational errors which were not considered in [NOCE]. Again, each step of an iteration has a computational errors and we take into account that these errors are, in general, different. An optimization problems with a composite objective function is studied in Chapter 7. A zero-sum game with two-players is considered in Chapter 8. A predicted decrease approximation-based method is used in Chapter 9 for constrained convex optimization. Chapter 10 is devoted to minimization of quasiconvex functions. Minimization of sharp weakly convex functions is discussed in Chapter 11. Chapter 12 is devoted to a generalized projected subgradient method for minimization of a convex function over a set which is not necessarily convex. The book is of interest for researchers and engineers working in optimization. It also can be useful in preparation courses for graduate students. The main feature of the book which appeals specifically to this audience is the study of the influence of computational errors for several important optimization algorithms. The book is of interest for experts in applications of optimization to engineering and economics.



Numerical Optimization With Computational Errors


Numerical Optimization With Computational Errors
DOWNLOAD
Author : Alexander J. Zaslavski
language : en
Publisher: Springer
Release Date : 2016-04-22

Numerical Optimization With Computational Errors written by Alexander J. Zaslavski and has been published by Springer this book supported file pdf, txt, epub, kindle and other format this book has been release on 2016-04-22 with Mathematics categories.


This book studies the approximate solutions of optimization problems in the presence of computational errors. A number of results are presented on the convergence behavior of algorithms in a Hilbert space; these algorithms are examined taking into account computational errors. The author illustrates that algorithms generate a good approximate solution, if computational errors are bounded from above by a small positive constant. Known computational errors are examined with the aim of determining an approximate solution. Researchers and students interested in the optimization theory and its applications will find this book instructive and informative. This monograph contains 16 chapters; including a chapters devoted to the subgradient projection algorithm, the mirror descent algorithm, gradient projection algorithm, the Weiszfelds method, constrained convex minimization problems, the convergence of a proximal point method in a Hilbert space, the continuous subgradient method, penalty methods and Newton’s method.



The Projected Subgradient Algorithm In Convex Optimization


The Projected Subgradient Algorithm In Convex Optimization
DOWNLOAD
Author : Alexander J. Zaslavski
language : en
Publisher: Springer Nature
Release Date : 2020-11-25

The Projected Subgradient Algorithm In Convex Optimization written by Alexander J. Zaslavski and has been published by Springer Nature this book supported file pdf, txt, epub, kindle and other format this book has been release on 2020-11-25 with Mathematics categories.


This focused monograph presents a study of subgradient algorithms for constrained minimization problems in a Hilbert space. The book is of interest for experts in applications of optimization to engineering and economics. The goal is to obtain a good approximate solution of the problem in the presence of computational errors. The discussion takes into consideration the fact that for every algorithm its iteration consists of several steps and that computational errors for different steps are different, in general. The book is especially useful for the reader because it contains solutions to a number of difficult and interesting problems in the numerical optimization. The subgradient projection algorithm is one of the most important tools in optimization theory and its applications. An optimization problem is described by an objective function and a set of feasible points. For this algorithm each iteration consists of two steps. The first step requires a calculation of a subgradient of the objective function; the second requires a calculation of a projection on the feasible set. The computational errors in each of these two steps are different. This book shows that the algorithm discussed, generates a good approximate solution, if all the computational errors are bounded from above by a small positive constant. Moreover, if computational errors for the two steps of the algorithm are known, one discovers an approximate solution and how many iterations one needs for this. In addition to their mathematical interest, the generalizations considered in this book have a significant practical meaning.



Convex Optimization


Convex Optimization
DOWNLOAD
Author : Stephen P. Boyd
language : en
Publisher: Cambridge University Press
Release Date : 2004-03-08

Convex Optimization written by Stephen P. Boyd and has been published by Cambridge University Press this book supported file pdf, txt, epub, kindle and other format this book has been release on 2004-03-08 with Business & Economics categories.


Convex optimization problems arise frequently in many different fields. This book provides a comprehensive introduction to the subject, and shows in detail how such problems can be solved numerically with great efficiency. The book begins with the basic elements of convex sets and functions, and then describes various classes of convex optimization problems. Duality and approximation techniques are then covered, as are statistical estimation techniques. Various geometrical problems are then presented, and there is detailed discussion of unconstrained and constrained minimization problems, and interior-point methods. The focus of the book is on recognizing convex optimization problems and then finding the most appropriate technique for solving them. It contains many worked examples and homework exercises and will appeal to students, researchers and practitioners in fields such as engineering, computer science, mathematics, statistics, finance and economics.



Lectures On Modern Convex Optimization


Lectures On Modern Convex Optimization
DOWNLOAD
Author : Aharon Ben-Tal
language : fr
Publisher: SIAM
Release Date : 2001-01-01

Lectures On Modern Convex Optimization written by Aharon Ben-Tal and has been published by SIAM this book supported file pdf, txt, epub, kindle and other format this book has been release on 2001-01-01 with Technology & Engineering categories.


Here is a book devoted to well-structured and thus efficiently solvable convex optimization problems, with emphasis on conic quadratic and semidefinite programming. The authors present the basic theory underlying these problems as well as their numerous applications in engineering, including synthesis of filters, Lyapunov stability analysis, and structural design. The authors also discuss the complexity issues and provide an overview of the basic theory of state-of-the-art polynomial time interior point methods for linear, conic quadratic, and semidefinite programming. The book's focus on well-structured convex problems in conic form allows for unified theoretical and algorithmical treatment of a wide spectrum of important optimization problems arising in applications.



Robust Optimization


Robust Optimization
DOWNLOAD
Author : Aharon Ben-Tal
language : en
Publisher: Princeton University Press
Release Date : 2009-08-10

Robust Optimization written by Aharon Ben-Tal and has been published by Princeton University Press this book supported file pdf, txt, epub, kindle and other format this book has been release on 2009-08-10 with Mathematics categories.


Robust optimization is still a relatively new approach to optimization problems affected by uncertainty, but it has already proved so useful in real applications that it is difficult to tackle such problems today without considering this powerful methodology. Written by the principal developers of robust optimization, and describing the main achievements of a decade of research, this is the first book to provide a comprehensive and up-to-date account of the subject. Robust optimization is designed to meet some major challenges associated with uncertainty-affected optimization problems: to operate under lack of full information on the nature of uncertainty; to model the problem in a form that can be solved efficiently; and to provide guarantees about the performance of the solution. The book starts with a relatively simple treatment of uncertain linear programming, proceeding with a deep analysis of the interconnections between the construction of appropriate uncertainty sets and the classical chance constraints (probabilistic) approach. It then develops the robust optimization theory for uncertain conic quadratic and semidefinite optimization problems and dynamic (multistage) problems. The theory is supported by numerous examples and computational illustrations. An essential book for anyone working on optimization and decision making under uncertainty, Robust Optimization also makes an ideal graduate textbook on the subject.



Optimization On Solution Sets Of Common Fixed Point Problems


Optimization On Solution Sets Of Common Fixed Point Problems
DOWNLOAD
Author : Alexander J. Zaslavski
language : en
Publisher: Springer Nature
Release Date : 2021-08-09

Optimization On Solution Sets Of Common Fixed Point Problems written by Alexander J. Zaslavski and has been published by Springer Nature this book supported file pdf, txt, epub, kindle and other format this book has been release on 2021-08-09 with Mathematics categories.


This book is devoted to a detailed study of the subgradient projection method and its variants for convex optimization problems over the solution sets of common fixed point problems and convex feasibility problems. These optimization problems are investigated to determine good solutions obtained by different versions of the subgradient projection algorithm in the presence of sufficiently small computational errors. The use of selected algorithms is highlighted including the Cimmino type subgradient, the iterative subgradient, and the dynamic string-averaging subgradient. All results presented are new. Optimization problems where the underlying constraints are the solution sets of other problems, frequently occur in applied mathematics. The reader should not miss the section in Chapter 1 which considers some examples arising in the real world applications. The problems discussed have an important impact in optimization theory as well. The book will be useful for researches interested in the optimization theory and its applications.



Uncertainty Quantification In Variational Inequalities


Uncertainty Quantification In Variational Inequalities
DOWNLOAD
Author : Joachim Gwinner
language : en
Publisher: CRC Press
Release Date : 2021-12-21

Uncertainty Quantification In Variational Inequalities written by Joachim Gwinner and has been published by CRC Press this book supported file pdf, txt, epub, kindle and other format this book has been release on 2021-12-21 with Mathematics categories.


Uncertainty Quantification (UQ) is an emerging and extremely active research discipline which aims to quantitatively treat any uncertainty in applied models. The primary objective of Uncertainty Quantification in Variational Inequalities: Theory, Numerics, and Applications is to present a comprehensive treatment of UQ in variational inequalities and some of its generalizations emerging from various network, economic, and engineering models. Some of the developed techniques also apply to machine learning, neural networks, and related fields. Features First book on UQ in variational inequalities emerging from various network, economic, and engineering models Completely self-contained and lucid in style Aimed for a diverse audience including applied mathematicians, engineers, economists, and professionals from academia Includes the most recent developments on the subject which so far have only been available in the research literature



Optimization In Banach Spaces


Optimization In Banach Spaces
DOWNLOAD
Author : Alexander J. Zaslavski
language : en
Publisher: Springer Nature
Release Date : 2022-09-29

Optimization In Banach Spaces written by Alexander J. Zaslavski and has been published by Springer Nature this book supported file pdf, txt, epub, kindle and other format this book has been release on 2022-09-29 with Mathematics categories.


The book is devoted to the study of constrained minimization problems on closed and convex sets in Banach spaces with a Frechet differentiable objective function. Such problems are well studied in a finite-dimensional space and in an infinite-dimensional Hilbert space. When the space is Hilbert there are many algorithms for solving optimization problems including the gradient projection algorithm which is one of the most important tools in the optimization theory, nonlinear analysis and their applications. An optimization problem is described by an objective function and a set of feasible points. For the gradient projection algorithm each iteration consists of two steps. The first step is a calculation of a gradient of the objective function while in the second one we calculate a projection on the feasible set. In each of these two steps there is a computational error. In our recent research we show that the gradient projection algorithm generates a good approximate solution, if all the computational errors are bounded from above by a small positive constant. It should be mentioned that the properties of a Hilbert space play an important role. When we consider an optimization problem in a general Banach space the situation becomes more difficult and less understood. On the other hand such problems arise in the approximation theory. The book is of interest for mathematicians working in optimization. It also can be useful in preparation courses for graduate students. The main feature of the book which appeals specifically to this audience is the study of algorithms for convex and nonconvex minimization problems in a general Banach space. The book is of interest for experts in applications of optimization to the approximation theory. In this book the goal is to obtain a good approximate solution of the constrained optimization problem in a general Banach space under the presence of computational errors. It is shown that the algorithm generates a good approximate solution, if the sequence of computational errors is bounded from above by a small constant. The book consists of four chapters. In the first we discuss several algorithms which are studied in the book and prove a convergence result for an unconstrained problem which is a prototype of our results for the constrained problem. In Chapter 2 we analyze convex optimization problems. Nonconvex optimization problems are studied in Chapter 3. In Chapter 4 we study continuous algorithms for minimization problems under the presence of computational errors. The algorithm generates a good approximate solution, if the sequence of computational errors is bounded from above by a small constant. The book consists of four chapters. In the first we discuss several algorithms which are studied in the book and prove a convergence result for an unconstrained problem which is a prototype of our results for the constrained problem. In Chapter 2 we analyze convex optimization problems. Nonconvex optimization problems are studied in Chapter 3. In Chapter 4 we study continuous algorithms for minimization problems under the presence of computational errors.



Introductory Lectures On Convex Optimization


Introductory Lectures On Convex Optimization
DOWNLOAD
Author : Y. Nesterov
language : en
Publisher: Springer Science & Business Media
Release Date : 2013-12-01

Introductory Lectures On Convex Optimization written by Y. Nesterov and has been published by Springer Science & Business Media this book supported file pdf, txt, epub, kindle and other format this book has been release on 2013-12-01 with Mathematics categories.


It was in the middle of the 1980s, when the seminal paper by Kar markar opened a new epoch in nonlinear optimization. The importance of this paper, containing a new polynomial-time algorithm for linear op timization problems, was not only in its complexity bound. At that time, the most surprising feature of this algorithm was that the theoretical pre diction of its high efficiency was supported by excellent computational results. This unusual fact dramatically changed the style and direc tions of the research in nonlinear optimization. Thereafter it became more and more common that the new methods were provided with a complexity analysis, which was considered a better justification of their efficiency than computational experiments. In a new rapidly develop ing field, which got the name "polynomial-time interior-point methods", such a justification was obligatory. Afteralmost fifteen years of intensive research, the main results of this development started to appear in monographs[12, 14, 16, 17, 18, 19]. Approximately at that time the author was asked to prepare a new course on nonlinear optimization for graduate students. The idea was to create a course which would reflect the new developments in the field. Actually, this was a major challenge. At the time only the theory of interior-point methods for linear optimization was polished enough to be explained to students. The general theory of self-concordant functions had appeared in print only once in the form of research monograph [12].