Minimization of Separable Convex Functions Subject to Equality and Box Constraints

Minimization of Separable Convex Functions Subject to Equality and Box Constraints PDF Author: Nainan Kovoor
Publisher:
ISBN:
Category : Algorithms
Languages : en
Pages : 7

Get Book Here

Book Description
Abstract: "We consider the problem of minimizing a separable differentiable strictly convex function on R[superscript n] subject to m equality constraints and upper and lower bounds (box constraints). We provide a parametric characterization in R[superscript m] of the family of solutions to this problem, thereby showing equivalence with a problem of search in an arrangement of hyperplanes in R[superscript m]. For the special case of the euclidean distance function and m = 2, we obtain a strongly polynomial algorithm running in time [theta](n2).

Minimization of Separable Convex Functions Subject to Equality and Box Constraints

Minimization of Separable Convex Functions Subject to Equality and Box Constraints PDF Author: Nainan Kovoor
Publisher:
ISBN:
Category : Algorithms
Languages : en
Pages : 7

Get Book Here

Book Description
Abstract: "We consider the problem of minimizing a separable differentiable strictly convex function on R[superscript n] subject to m equality constraints and upper and lower bounds (box constraints). We provide a parametric characterization in R[superscript m] of the family of solutions to this problem, thereby showing equivalence with a problem of search in an arrangement of hyperplanes in R[superscript m]. For the special case of the euclidean distance function and m = 2, we obtain a strongly polynomial algorithm running in time [theta](n2).

Convex Optimization

Convex Optimization PDF Author: Stephen P. Boyd
Publisher: Cambridge University Press
ISBN: 9780521833783
Category : Business & Economics
Languages : en
Pages : 744

Get Book Here

Book Description
Convex optimization problems arise frequently in many different fields. This book provides a comprehensive introduction to the subject, and shows in detail how such problems can be solved numerically with great efficiency. The book begins with the basic elements of convex sets and functions, and then describes various classes of convex optimization problems. Duality and approximation techniques are then covered, as are statistical estimation techniques. Various geometrical problems are then presented, and there is detailed discussion of unconstrained and constrained minimization problems, and interior-point methods. The focus of the book is on recognizing convex optimization problems and then finding the most appropriate technique for solving them. It contains many worked examples and homework exercises and will appeal to students, researchers and practitioners in fields such as engineering, computer science, mathematics, statistics, finance and economics.

Separable Optimization

Separable Optimization PDF Author: Stefan M. Stefanov
Publisher: Springer Nature
ISBN: 3030784010
Category : Mathematics
Languages : en
Pages : 360

Get Book Here

Book Description
In this book, the theory, methods and applications of separable optimization are considered. Some general results are presented, techniques of approximating the separable problem by linear programming problem, and dynamic programming are also studied. Convex separable programs subject to inequality/ equality constraint(s) and bounds on variables are also studied and convergent iterative algorithms of polynomial complexity are proposed. As an application, these algorithms are used in the implementation of stochastic quasigradient methods to some separable stochastic programs. The problems of numerical approximation of tabulated functions and numerical solution of overdetermined systems of linear algebraic equations and some systems of nonlinear equations are solved by separable convex unconstrained minimization problems. Some properties of the Knapsack polytope are also studied. This second edition includes a substantial amount of new and revised content. Three new chapters, 15-17, are included. Chapters 15-16 are devoted to the further analysis of the Knapsack problem. Chapter 17 is focused on the analysis of a nonlinear transportation problem. Three new Appendices (E-G) are also added to this edition and present technical details that help round out the coverage. Optimization problems and methods for solving the problems considered are interesting not only from the viewpoint of optimization theory, optimization methods and their applications, but also from the viewpoint of other fields of science, especially the artificial intelligence and machine learning fields within computer science. This book is intended for the researcher, practitioner, or engineer who is interested in the detailed treatment of separable programming and wants to take advantage of the latest theoretical and algorithmic results. It may also be used as a textbook for a special topics course or as a supplementary textbook for graduate courses on nonlinear and convex optimization.

Explicit Solutions for a General Class of Optimal Allocation Problems

Explicit Solutions for a General Class of Optimal Allocation Problems PDF Author:
Publisher:
ISBN:
Category :
Languages : en
Pages :

Get Book Here

Book Description


An Algorithm for Separable Convex Programming Under Linear Equality Constraints

An Algorithm for Separable Convex Programming Under Linear Equality Constraints PDF Author: James E. FALK
Publisher:
ISBN:
Category :
Languages : en
Pages : 23

Get Book Here

Book Description


Convex Analysis and Optimization

Convex Analysis and Optimization PDF Author: Dimitri Bertsekas
Publisher: Athena Scientific
ISBN: 1886529450
Category : Mathematics
Languages : en
Pages : 560

Get Book Here

Book Description
A uniquely pedagogical, insightful, and rigorous treatment of the analytical/geometrical foundations of optimization. The book provides a comprehensive development of convexity theory, and its rich applications in optimization, including duality, minimax/saddle point theory, Lagrange multipliers, and Lagrangian relaxation/nondifferentiable optimization. It is an excellent supplement to several of our books: Convex Optimization Theory (Athena Scientific, 2009), Convex Optimization Algorithms (Athena Scientific, 2015), Nonlinear Programming (Athena Scientific, 2016), Network Optimization (Athena Scientific, 1998), and Introduction to Linear Optimization (Athena Scientific, 1997). Aside from a thorough account of convex analysis and optimization, the book aims to restructure the theory of the subject, by introducing several novel unifying lines of analysis, including: 1) A unified development of minimax theory and constrained optimization duality as special cases of duality between two simple geometrical problems. 2) A unified development of conditions for existence of solutions of convex optimization problems, conditions for the minimax equality to hold, and conditions for the absence of a duality gap in constrained optimization. 3) A unification of the major constraint qualifications allowing the use of Lagrange multipliers for nonconvex constrained optimization, using the notion of constraint pseudonormality and an enhanced form of the Fritz John necessary optimality conditions. Among its features the book: a) Develops rigorously and comprehensively the theory of convex sets and functions, in the classical tradition of Fenchel and Rockafellar b) Provides a geometric, highly visual treatment of convex and nonconvex optimization problems, including existence of solutions, optimality conditions, Lagrange multipliers, and duality c) Includes an insightful and comprehensive presentation of minimax theory and zero sum games, and its connection with duality d) Describes dual optimization, the associated computational methods, including the novel incremental subgradient methods, and applications in linear, quadratic, and integer programming e) Contains many examples, illustrations, and exercises with complete solutions (about 200 pages) posted at the publisher's web site http://www.athenasc.com/convexity.html

Convex Optimization & Euclidean Distance Geometry

Convex Optimization & Euclidean Distance Geometry PDF Author: Jon Dattorro
Publisher: Meboo Publishing USA
ISBN: 0976401304
Category : Mathematics
Languages : en
Pages : 776

Get Book Here

Book Description
The study of Euclidean distance matrices (EDMs) fundamentally asks what can be known geometrically given onlydistance information between points in Euclidean space. Each point may represent simply locationor, abstractly, any entity expressible as a vector in finite-dimensional Euclidean space.The answer to the question posed is that very much can be known about the points;the mathematics of this combined study of geometry and optimization is rich and deep.Throughout we cite beacons of historical accomplishment.The application of EDMs has already proven invaluable in discerning biological molecular conformation.The emerging practice of localization in wireless sensor networks, the global positioning system (GPS), and distance-based pattern recognitionwill certainly simplify and benefit from this theory.We study the pervasive convex Euclidean bodies and their various representations.In particular, we make convex polyhedra, cones, and dual cones more visceral through illustration, andwe study the geometric relation of polyhedral cones to nonorthogonal bases biorthogonal expansion.We explain conversion between halfspace- and vertex-descriptions of convex cones,we provide formulae for determining dual cones,and we show how classic alternative systems of linear inequalities or linear matrix inequalities and optimality conditions can be explained by generalized inequalities in terms of convex cones and their duals.The conic analogue to linear independence, called conic independence, is introducedas a new tool in the study of classical cone theory; the logical next step in the progression:linear, affine, conic.Any convex optimization problem has geometric interpretation.This is a powerful attraction: the ability to visualize geometry of an optimization problem.We provide tools to make visualization easier.The concept of faces, extreme points, and extreme directions of convex Euclidean bodiesis explained here, crucial to understanding convex optimization.The convex cone of positive semidefinite matrices, in particular, is studied in depth.We mathematically interpret, for example,its inverse image under affine transformation, and we explainhow higher-rank subsets of its boundary united with its interior are convex.The Chapter on "Geometry of convex functions",observes analogies between convex sets and functions:The set of all vector-valued convex functions is a closed convex cone.Included among the examples in this chapter, we show how the real affinefunction relates to convex functions as the hyperplane relates to convex sets.Here, also, pertinent results formultidimensional convex functions are presented that are largely ignored in the literature;tricks and tips for determining their convexityand discerning their geometry, particularly with regard to matrix calculus which remains largely unsystematizedwhen compared with the traditional practice of ordinary calculus.Consequently, we collect some results of matrix differentiation in the appendices.The Euclidean distance matrix (EDM) is studied,its properties and relationship to both positive semidefinite and Gram matrices.We relate the EDM to the four classical axioms of the Euclidean metric;thereby, observing the existence of an infinity of axioms of the Euclidean metric beyondthe triangle inequality. We proceed byderiving the fifth Euclidean axiom and then explain why furthering this endeavoris inefficient because the ensuing criteria (while describing polyhedra)grow linearly in complexity and number.Some geometrical problems solvable via EDMs,EDM problems posed as convex optimization, and methods of solution arepresented;\eg, we generate a recognizable isotonic map of the United States usingonly comparative distance information (no distance information, only distance inequalities).We offer a new proof of the classic Schoenberg criterion, that determines whether a candidate matrix is an EDM. Our proofrelies on fundamental geometry; assuming, any EDM must correspond to a list of points contained in some polyhedron(possibly at its vertices) and vice versa.It is not widely known that the Schoenberg criterion implies nonnegativity of the EDM entries; proved here.We characterize the eigenvalues of an EDM matrix and then devisea polyhedral cone required for determining membership of a candidate matrix(in Cayley-Menger form) to the convex cone of Euclidean distance matrices (EDM cone); \ie,a candidate is an EDM if and only if its eigenspectrum belongs to a spectral cone for EDM^N.We will see spectral cones are not unique.In the chapter "EDM cone", we explain the geometric relationship betweenthe EDM cone, two positive semidefinite cones, and the elliptope.We illustrate geometric requirements, in particular, for projection of a candidate matrixon a positive semidefinite cone that establish its membership to the EDM cone. The faces of the EDM cone are described,but still open is the question whether all its faces are exposed as they are for the positive semidefinite cone.The classic Schoenberg criterion, relating EDM and positive semidefinite cones, isrevealed to be a discretized membership relation (a generalized inequality, a new Farkas''''''''-like lemma)between the EDM cone and its ordinary dual. A matrix criterion for membership to the dual EDM cone is derived thatis simpler than the Schoenberg criterion.We derive a new concise expression for the EDM cone and its dual involvingtwo subspaces and a positive semidefinite cone."Semidefinite programming" is reviewedwith particular attention to optimality conditionsof prototypical primal and dual conic programs,their interplay, and the perturbation method of rank reduction of optimal solutions(extant but not well-known).We show how to solve a ubiquitous platonic combinatorial optimization problem from linear algebra(the optimal Boolean solution x to Ax=b)via semidefinite program relaxation.A three-dimensional polyhedral analogue for the positive semidefinite cone of 3X3 symmetricmatrices is introduced; a tool for visualizing in 6 dimensions.In "EDM proximity"we explore methods of solution to a few fundamental and prevalentEuclidean distance matrix proximity problems; the problem of finding that Euclidean distance matrix closestto a given matrix in the Euclidean sense.We pay particular attention to the problem when compounded with rank minimization.We offer a new geometrical proof of a famous result discovered by Eckart \& Young in 1936 regarding Euclideanprojection of a point on a subset of the positive semidefinite cone comprising all positive semidefinite matriceshaving rank not exceeding a prescribed limit rho.We explain how this problem is transformed to a convex optimization for any rank rho.

Convex Analysis and Minimization Algorithms II

Convex Analysis and Minimization Algorithms II PDF Author: Jean-Baptiste Hiriart-Urruty
Publisher: Springer Science & Business Media
ISBN: 9783540568520
Category : Business & Economics
Languages : en
Pages : 374

Get Book Here

Book Description
From the reviews: "The account is quite detailed and is written in a manner that will appeal to analysts and numerical practitioners alike...they contain everything from rigorous proofs to tables of numerical calculations.... one of the strong features of these books...that they are designed not for the expert, but for those who whish to learn the subject matter starting from little or no background...there are numerous examples, and counter-examples, to back up the theory...To my knowledge, no other authors have given such a clear geometric account of convex analysis." "This innovative text is well written, copiously illustrated, and accessible to a wide audience"

Linear Algebra and Optimization for Machine Learning

Linear Algebra and Optimization for Machine Learning PDF Author: Charu C. Aggarwal
Publisher: Springer Nature
ISBN: 3030403440
Category : Computers
Languages : en
Pages : 507

Get Book Here

Book Description
This textbook introduces linear algebra and optimization in the context of machine learning. Examples and exercises are provided throughout the book. A solution manual for the exercises at the end of each chapter is available to teaching instructors. This textbook targets graduate level students and professors in computer science, mathematics and data science. Advanced undergraduate students can also use this textbook. The chapters for this textbook are organized as follows: 1. Linear algebra and its applications: The chapters focus on the basics of linear algebra together with their common applications to singular value decomposition, matrix factorization, similarity matrices (kernel methods), and graph analysis. Numerous machine learning applications have been used as examples, such as spectral clustering, kernel-based classification, and outlier detection. The tight integration of linear algebra methods with examples from machine learning differentiates this book from generic volumes on linear algebra. The focus is clearly on the most relevant aspects of linear algebra for machine learning and to teach readers how to apply these concepts. 2. Optimization and its applications: Much of machine learning is posed as an optimization problem in which we try to maximize the accuracy of regression and classification models. The “parent problem” of optimization-centric machine learning is least-squares regression. Interestingly, this problem arises in both linear algebra and optimization, and is one of the key connecting problems of the two fields. Least-squares regression is also the starting point for support vector machines, logistic regression, and recommender systems. Furthermore, the methods for dimensionality reduction and matrix factorization also require the development of optimization methods. A general view of optimization in computational graphs is discussed together with its applications to back propagation in neural networks. A frequent challenge faced by beginners in machine learning is the extensive background required in linear algebra and optimization. One problem is that the existing linear algebra and optimization courses are not specific to machine learning; therefore, one would typically have to complete more course material than is necessary to pick up machine learning. Furthermore, certain types of ideas and tricks from optimization and linear algebra recur more frequently in machine learning than other application-centric settings. Therefore, there is significant value in developing a view of linear algebra and optimization that is better suited to the specific perspective of machine learning.

Convex Analysis and Minimization Algorithms I

Convex Analysis and Minimization Algorithms I PDF Author: Jean-Baptiste Hiriart-Urruty
Publisher: Springer Science & Business Media
ISBN: 3662027968
Category : Mathematics
Languages : en
Pages : 432

Get Book Here

Book Description
Convex Analysis may be considered as a refinement of standard calculus, with equalities and approximations replaced by inequalities. As such, it can easily be integrated into a graduate study curriculum. Minimization algorithms, more specifically those adapted to non-differentiable functions, provide an immediate application of convex analysis to various fields related to optimization and operations research. These two topics making up the title of the book, reflect the two origins of the authors, who belong respectively to the academic world and to that of applications. Part I can be used as an introductory textbook (as a basis for courses, or for self-study); Part II continues this at a higher technical level and is addressed more to specialists, collecting results that so far have not appeared in books.