fbpx
Wikipedia

Duality (optimization)

In mathematical optimization theory, duality or the duality principle is the principle that optimization problems may be viewed from either of two perspectives, the primal problem or the dual problem. If the primal is a minimization problem then the dual is a maximization problem (and vice versa). Any feasible solution to the primal (minimization) problem is at least as large as any feasible solution to the dual (maximization) problem. Therefore, the solution to the primal is an upper bound to the solution of the dual, and the solution of the dual is a lower bound to the solution of the primal.[1] This fact is called weak duality.

In general, the optimal values of the primal and dual problems need not be equal. Their difference is called the duality gap. For convex optimization problems, the duality gap is zero under a constraint qualification condition. This fact is called strong duality.

Dual problem edit

Usually the term "dual problem" refers to the Lagrangian dual problem but other dual problems are used – for example, the Wolfe dual problem and the Fenchel dual problem. The Lagrangian dual problem is obtained by forming the Lagrangian of a minimization problem by using nonnegative Lagrange multipliers to add the constraints to the objective function, and then solving for the primal variable values that minimize the original objective function. This solution gives the primal variables as functions of the Lagrange multipliers, which are called dual variables, so that the new problem is to maximize the objective function with respect to the dual variables under the derived constraints on the dual variables (including at least the nonnegativity constraints).

In general given two dual pairs of separated locally convex spaces   and   and the function  , we can define the primal problem as finding   such that   In other words, if   exists,   is the minimum of the function   and the infimum (greatest lower bound) of the function is attained.

If there are constraint conditions, these can be built into the function   by letting   where   is a suitable function on   that has a minimum 0 on the constraints, and for which one can prove that  . The latter condition is trivially, but not always conveniently, satisfied for the characteristic function (i.e.   for   satisfying the constraints and   otherwise). Then extend   to a perturbation function   such that  .[2]

The duality gap is the difference of the right and left hand sides of the inequality

 

where   is the convex conjugate in both variables and   denotes the supremum (least upper bound).[2][3][4]

Duality gap edit

The duality gap is the difference between the values of any primal solutions and any dual solutions. If   is the optimal dual value and   is the optimal primal value, then the duality gap is equal to  . This value is always greater than or equal to 0 (for minimization problems). The duality gap is zero if and only if strong duality holds. Otherwise the gap is strictly positive and weak duality holds.[5]

In computational optimization, another "duality gap" is often reported, which is the difference in value between any dual solution and the value of a feasible but suboptimal iterate for the primal problem. This alternative "duality gap" quantifies the discrepancy between the value of a current feasible but suboptimal iterate for the primal problem and the value of the dual problem; the value of the dual problem is, under regularity conditions, equal to the value of the convex relaxation of the primal problem: The convex relaxation is the problem arising replacing a non-convex feasible set with its closed convex hull and with replacing a non-convex function with its convex closure, that is the function that has the epigraph that is the closed convex hull of the original primal objective function.[6][7][8][9][10][11][12][13][14][15][16]

Linear case edit

Linear programming problems are optimization problems in which the objective function and the constraints are all linear. In the primal problem, the objective function is a linear combination of n variables. There are m constraints, each of which places an upper bound on a linear combination of the n variables. The goal is to maximize the value of the objective function subject to the constraints. A solution is a vector (a list) of n values that achieves the maximum value for the objective function.

In the dual problem, the objective function is a linear combination of the m values that are the limits in the m constraints from the primal problem. There are n dual constraints, each of which places a lower bound on a linear combination of m dual variables.

Relationship between the primal problem and the dual problem edit

In the linear case, in the primal problem, from each sub-optimal point that satisfies all the constraints, there is a direction or subspace of directions to move that increases the objective function. Moving in any such direction is said to remove slack between the candidate solution and one or more constraints. An infeasible value of the candidate solution is one that exceeds one or more of the constraints.

In the dual problem, the dual vector multiplies the constraints that determine the positions of the constraints in the primal. Varying the dual vector in the dual problem is equivalent to revising the upper bounds in the primal problem. The lowest upper bound is sought. That is, the dual vector is minimized in order to remove slack between the candidate positions of the constraints and the actual optimum. An infeasible value of the dual vector is one that is too low. It sets the candidate positions of one or more of the constraints in a position that excludes the actual optimum.

This intuition is made formal by the equations in Linear programming: Duality.

Nonlinear case edit

In nonlinear programming, the constraints are not necessarily linear. Nonetheless, many of the same principles apply.

To ensure that the global maximum of a non-linear problem can be identified easily, the problem formulation often requires that the functions be convex and have compact lower level sets. This is the significance of the Karush–Kuhn–Tucker conditions. They provide necessary conditions for identifying local optima of non-linear programming problems. There are additional conditions (constraint qualifications) that are necessary so that it will be possible to define the direction to an optimal solution. An optimal solution is one that is a local optimum, but possibly not a global optimum.

Lagrange duality edit

Motivation.[17] Suppose we want to solve the following nonlinear programming problem:

 

The problem has constraints; we would like to convert it to a program without constraints. Theoretically, it is possible to do it by minimizing the function J(x), defined as

 

where I is an infinite step function: I[u]=0 if u≤0, and I[u]=∞ otherwise. But J(x) is hard to solve as it is not continuous. It is possible to "approximate" I[u] by λu, where λ is a positive constant. This yields a function known as the lagrangian:

 

Note that, for every x,

 .

Proof:

  • If x satisfies all constraints fi(x)≤0, then L(x,λ) is maximized when taking λ=0, and its value is then f(x);
  • If x violates some constraint, fi(x)>0 for some i, then L(x,λ)→∞ when λi→∞.

Therefore, the original problem is equivalent to:

 .

By reversing the order of min and max, we get:

 .

The dual function is the inner problem in the above formula:

 .

The Lagrangian dual program is the program of maximizing g:

 .

The optimal solution to the dual program is a lower bound for the optimal solution of the original (primal) program; this is the weak duality principle. If the primal problem is convex and bounded from below, and there exists a point in which all nonlinear constraints are strictly satisfied (Slater's condition), then the optimal solution to the dual program equals the optimal solution of the primal program; this is the strong duality principle. In this case, we can solve the primal program by finding an optimal solution λ* to the dual program, and then solving:

 .

Note that, to use either the weak or the strong duality principle, we need a way to compute g(λ). In general this may be hard, as we need to solve a different minimization problem for every λ. But for some classes of functions, it is possible to get an explicit formula for g(). Solving the primal and dual programs together is often easier than solving only one of them. Examples are linear programming and quadratic programming. A better and more general approach to duality is provided by Fenchel's duality theorem.[18]: Sub.3.3.1 

Another condition in which the min-max and max-min are equal is when the Lagrangian has a saddle point: (x∗, λ∗) is a saddle point of the Lagrange function L if and only if x∗ is an optimal solution to the primal, λ∗ is an optimal solution to the dual, and the optimal values in the indicated problems are equal to each other.[18]: Prop.3.2.2 

The strong Lagrange principle edit

Given a nonlinear programming problem in standard form

 

with the domain   having non-empty interior, the Lagrangian function   is defined as

 

The vectors   and   are called the dual variables or Lagrange multiplier vectors associated with the problem. The Lagrange dual function   is defined as

 

The dual function g is concave, even when the initial problem is not convex, because it is a point-wise infimum of affine functions. The dual function yields lower bounds on the optimal value   of the initial problem; for any   and any   we have  .

If a constraint qualification such as Slater's condition holds and the original problem is convex, then we have strong duality, i.e.  .

Convex problems edit

For a convex minimization problem with inequality constraints,

 

the Lagrangian dual problem is

 

where the objective function is the Lagrange dual function. Provided that the functions   and   are continuously differentiable, the infimum occurs where the gradient is equal to zero. The problem

 

is called the Wolfe dual problem. This problem may be difficult to deal with computationally, because the objective function is not concave in the joint variables  . Also, the equality constraint   is nonlinear in general, so the Wolfe dual problem is typically a nonconvex optimization problem. In any case, weak duality holds.[19]

History edit

According to George Dantzig, the duality theorem for linear optimization was conjectured by John von Neumann immediately after Dantzig presented the linear programming problem. Von Neumann noted that he was using information from his game theory, and conjectured that two person zero sum matrix game was equivalent to linear programming. Rigorous proofs were first published in 1948 by Albert W. Tucker and his group. (Dantzig's foreword to Nering and Tucker, 1993)

Applications edit

In support vector machines (SVMs), the formulating the primal problem of SVMs as the dual problem can be used to implement Kernel trick, but the latter has higher time complexity in the historical cases.

See also edit

Notes edit

  1. ^ Boyd, Stephen P.; Vandenberghe, Lieven (2004). Convex Optimization (pdf). Cambridge University Press. p. 216. ISBN 978-0-521-83378-3. Retrieved October 15, 2011.
  2. ^ a b Boţ, Radu Ioan; Wanka, Gert; Grad, Sorin-Mihai (2009). Duality in Vector Optimization. Springer. ISBN 978-3-642-02885-4.
  3. ^ Csetnek, Ernö Robert (2010). Overcoming the failure of the classical generalized interior-point regularity conditions in convex optimization. Applications of the duality theory to enlargements of maximal monotone operators. Logos Verlag Berlin GmbH. ISBN 978-3-8325-2503-3.
  4. ^ Zălinescu, Constantin (2002). Convex analysis in general vector spaces. River Edge, NJ: World Scientific Publishing Co., Inc. pp. 106–113. ISBN 981-238-067-1. MR 1921556.
  5. ^ Borwein, Jonathan; Zhu, Qiji (2005). Techniques of Variational Analysis. Springer. ISBN 978-1-4419-2026-3.
  6. ^ Ahuja, Ravindra K.; Magnanti, Thomas L.; Orlin, James B. (1993). Network Flows: Theory, Algorithms and Applications. Prentice Hall. ISBN 0-13-617549-X.
  7. ^ Bertsekas, Dimitri; Nedic, Angelia; Ozdaglar, Asuman (2003). Convex Analysis and Optimization. Athena Scientific. ISBN 1-886529-45-0.
  8. ^ Bertsekas, Dimitri P. (1999). Nonlinear Programming (2nd ed.). Athena Scientific. ISBN 1-886529-00-0.
  9. ^ Bertsekas, Dimitri P. (2009). Convex Optimization Theory. Athena Scientific. ISBN 978-1-886529-31-1.
  10. ^ Bonnans, J. Frédéric; Gilbert, J. Charles; Lemaréchal, Claude; Sagastizábal, Claudia A. (2006). Numerical optimization: Theoretical and practical aspects. Universitext (Second revised ed. of translation of 1997 French ed.). Berlin: Springer-Verlag. pp. xiv+490. doi:10.1007/978-3-540-35447-5. ISBN 3-540-35445-X. MR 2265882.
  11. ^ Hiriart-Urruty, Jean-Baptiste; Lemaréchal, Claude (1993). Convex analysis and minimization algorithms, Volume I: Fundamentals. Grundlehren der Mathematischen Wissenschaften [Fundamental Principles of Mathematical Sciences]. Vol. 305. Berlin: Springer-Verlag. pp. xviii+417. ISBN 3-540-56850-6. MR 1261420.
  12. ^ Hiriart-Urruty, Jean-Baptiste; Lemaréchal, Claude (1993). "14 Duality for Practitioners". Convex analysis and minimization algorithms, Volume II: Advanced theory and bundle methods. Grundlehren der Mathematischen Wissenschaften [Fundamental Principles of Mathematical Sciences]. Vol. 306. Berlin: Springer-Verlag. pp. xviii+346. ISBN 3-540-56852-2. MR 1295240.
  13. ^ Lasdon, Leon S. (2002) [Reprint of the 1970 Macmillan]. Optimization theory for large systems. Mineola, New York: Dover Publications, Inc. pp. xiii+523. ISBN 978-0-486-41999-2. MR 1888251.
  14. ^ Lemaréchal, Claude (2001). "Lagrangian relaxation". In Jünger, Michael; Naddef, Denis (eds.). Computational combinatorial optimization: Papers from the Spring School held in Schloß Dagstuhl, May 15–19, 2000. Lecture Notes in Computer Science (LNCS). Vol. 2241. Berlin: Springer-Verlag. pp. 112–156. doi:10.1007/3-540-45586-8_4. ISBN 3-540-42877-1. MR 1900016. S2CID 9048698.
  15. ^ Minoux, Michel (1986). Mathematical programming: Theory and algorithms. Egon Balas (forward); Steven Vajda (trans) from the (1983 Paris: Dunod) French. Chichester: A Wiley-Interscience Publication. John Wiley & Sons, Ltd. pp. xxviii+489. ISBN 0-471-90170-9. MR 0868279. (2008 Second ed., in French: Programmation mathématique : Théorie et algorithmes, Éditions Tec & Doc, Paris, 2008. xxx+711 pp. ).
  16. ^ Shapiro, Jeremy F. (1979). Mathematical programming: Structures and algorithms. New York: Wiley-Interscience [John Wiley & Sons]. pp. xvi+388. ISBN 0-471-77886-9. MR 0544669.
  17. ^ David Knowles (2010). "Lagrangian Duality for Dummies" (PDF).
  18. ^ a b Nemirovsky and Ben-Tal (2023). "Optimization III: Convex Optimization" (PDF).
  19. ^ Geoffrion, Arthur M. (1971). "Duality in Nonlinear Programming: A Simplified Applications-Oriented Development". SIAM Review. 13 (1): 1–37. doi:10.1137/1013001. JSTOR 2028848.

References edit

Books edit

  • Ahuja, Ravindra K.; Magnanti, Thomas L.; Orlin, James B. (1993). Network Flows: Theory, Algorithms and Applications. Prentice Hall. ISBN 0-13-617549-X.
  • Bertsekas, Dimitri; Nedic, Angelia; Ozdaglar, Asuman (2003). Convex Analysis and Optimization. Athena Scientific. ISBN 1-886529-45-0.
  • Bertsekas, Dimitri P. (1999). Nonlinear Programming (2nd ed.). Athena Scientific. ISBN 1-886529-00-0.
  • Bertsekas, Dimitri P. (2009). Convex Optimization Theory. Athena Scientific. ISBN 978-1-886529-31-1.
  • Bonnans, J. Frédéric; Gilbert, J. Charles; Lemaréchal, Claude; Sagastizábal, Claudia A. (2006). Numerical optimization: Theoretical and practical aspects. Universitext (Second revised ed. of translation of 1997 French ed.). Berlin: Springer-Verlag. pp. xiv+490. doi:10.1007/978-3-540-35447-5. ISBN 3-540-35445-X. MR 2265882.
  • Cook, William J.; Cunningham, William H.; Pulleyblank, William R.; Schrijver, Alexander (November 12, 1997). Combinatorial Optimization (1st ed.). John Wiley & Sons. ISBN 0-471-55894-X.
  • Dantzig, George B. (1963). Linear Programming and Extensions. Princeton, NJ: Princeton University Press.
  • Hiriart-Urruty, Jean-Baptiste; Lemaréchal, Claude (1993). Convex analysis and minimization algorithms, Volume I: Fundamentals. Grundlehren der Mathematischen Wissenschaften [Fundamental Principles of Mathematical Sciences]. Vol. 305. Berlin: Springer-Verlag. pp. xviii+417. ISBN 3-540-56850-6. MR 1261420.
  • Hiriart-Urruty, Jean-Baptiste; Lemaréchal, Claude (1993). "14 Duality for Practitioners". Convex analysis and minimization algorithms, Volume II: Advanced theory and bundle methods. Grundlehren der Mathematischen Wissenschaften [Fundamental Principles of Mathematical Sciences]. Vol. 306. Berlin: Springer-Verlag. pp. xviii+346. ISBN 3-540-56852-2. MR 1295240.
  • Lasdon, Leon S. (2002) [Reprint of the 1970 Macmillan]. Optimization theory for large systems. Mineola, New York: Dover Publications, Inc. pp. xiii+523. ISBN 978-0-486-41999-2. MR 1888251.
  • Lawler, Eugene (2001). "4.5. Combinatorial Implications of Max-Flow Min-Cut Theorem, 4.6. Linear Programming Interpretation of Max-Flow Min-Cut Theorem". Combinatorial Optimization: Networks and Matroids. Dover. pp. 117–120. ISBN 0-486-41453-1.
  • Lemaréchal, Claude (2001). "Lagrangian relaxation". In Jünger, Michael; Naddef, Denis (eds.). Computational combinatorial optimization: Papers from the Spring School held in Schloß Dagstuhl, May 15–19, 2000. Lecture Notes in Computer Science (LNCS). Vol. 2241. Berlin: Springer-Verlag. pp. 112–156. doi:10.1007/3-540-45586-8_4. ISBN 3-540-42877-1. MR 1900016. S2CID 9048698.
  • Minoux, Michel (1986). Mathematical programming: Theory and algorithms. Egon Balas (forward); Steven Vajda (trans) from the (1983 Paris: Dunod) French. Chichester: A Wiley-Interscience Publication. John Wiley & Sons, Ltd. pp. xxviii+489. ISBN 0-471-90170-9. MR 0868279. (2008 Second ed., in French: Programmation mathématique : Théorie et algorithmes, Éditions Tec & Doc, Paris, 2008. xxx+711 pp. )).
  • Nering, Evar D.; Tucker, Albert W. (1993). Linear Programming and Related Problems. Boston, MA: Academic Press. ISBN 978-0-12-515440-6.
  • Papadimitriou, Christos H.; Steiglitz, Kenneth (July 1998). Combinatorial Optimization: Algorithms and Complexity (Unabridged ed.). Dover. ISBN 0-486-40258-4.
  • Ruszczyński, Andrzej (2006). Nonlinear Optimization. Princeton, NJ: Princeton University Press. pp. xii+454. ISBN 978-0-691-11915-1. MR 2199043.

Articles edit

  • Everett, Hugh III (1963). . Operations Research. 11 (3): 399–417. doi:10.1287/opre.11.3.399. JSTOR 168028. MR 0152360. Archived from the original on 2011-07-24.
  • Kiwiel, Krzysztof C.; Larsson, Torbjörn; Lindberg, P. O. (August 2007). . Mathematics of Operations Research. 32 (3): 669–686. doi:10.1287/moor.1070.0261. MR 2348241. Archived from the original on 2011-07-26. Retrieved 2011-05-12.
  • Duality in Linear Programming Gary D. Knott

duality, optimization, mathematical, optimization, theory, duality, duality, principle, principle, that, optimization, problems, viewed, from, either, perspectives, primal, problem, dual, problem, primal, minimization, problem, then, dual, maximization, proble. In mathematical optimization theory duality or the duality principle is the principle that optimization problems may be viewed from either of two perspectives the primal problem or the dual problem If the primal is a minimization problem then the dual is a maximization problem and vice versa Any feasible solution to the primal minimization problem is at least as large as any feasible solution to the dual maximization problem Therefore the solution to the primal is an upper bound to the solution of the dual and the solution of the dual is a lower bound to the solution of the primal 1 This fact is called weak duality In general the optimal values of the primal and dual problems need not be equal Their difference is called the duality gap For convex optimization problems the duality gap is zero under a constraint qualification condition This fact is called strong duality Contents 1 Dual problem 1 1 Duality gap 2 Linear case 2 1 Relationship between the primal problem and the dual problem 3 Nonlinear case 3 1 Lagrange duality 3 2 The strong Lagrange principle 3 3 Convex problems 4 History 5 Applications 6 See also 7 Notes 8 References 8 1 Books 8 2 ArticlesDual problem editUsually the term dual problem refers to the Lagrangian dual problem but other dual problems are used for example the Wolfe dual problem and the Fenchel dual problem The Lagrangian dual problem is obtained by forming the Lagrangian of a minimization problem by using nonnegative Lagrange multipliers to add the constraints to the objective function and then solving for the primal variable values that minimize the original objective function This solution gives the primal variables as functions of the Lagrange multipliers which are called dual variables so that the new problem is to maximize the objective function with respect to the dual variables under the derived constraints on the dual variables including at least the nonnegativity constraints In general given two dual pairs of separated locally convex spaces X X displaystyle left X X right nbsp and Y Y displaystyle left Y Y right nbsp and the function f X R displaystyle f X to mathbb R cup infty nbsp we can define the primal problem as finding x displaystyle hat x nbsp such that f x inf x X f x displaystyle f hat x inf x in X f x nbsp In other words if x displaystyle hat x nbsp exists f x displaystyle f hat x nbsp is the minimum of the function f displaystyle f nbsp and the infimum greatest lower bound of the function is attained If there are constraint conditions these can be built into the function f displaystyle f nbsp by letting f f I c o n s t r a i n t s displaystyle tilde f f I mathrm constraints nbsp where I c o n s t r a i n t s displaystyle I mathrm constraints nbsp is a suitable function on X displaystyle X nbsp that has a minimum 0 on the constraints and for which one can prove that inf x X f x inf x c o n s t r a i n e d f x displaystyle inf x in X tilde f x inf x mathrm constrained f x nbsp The latter condition is trivially but not always conveniently satisfied for the characteristic function i e I c o n s t r a i n t s x 0 displaystyle I mathrm constraints x 0 nbsp for x displaystyle x nbsp satisfying the constraints and I c o n s t r a i n t s x displaystyle I mathrm constraints x infty nbsp otherwise Then extend f displaystyle tilde f nbsp to a perturbation function F X Y R displaystyle F X times Y to mathbb R cup infty nbsp such that F x 0 f x displaystyle F x 0 tilde f x nbsp 2 The duality gap is the difference of the right and left hand sides of the inequality sup y Y F 0 y inf x X F x 0 displaystyle sup y in Y F 0 y leq inf x in X F x 0 nbsp where F displaystyle F nbsp is the convex conjugate in both variables and sup displaystyle sup nbsp denotes the supremum least upper bound 2 3 4 Duality gap edit Main article Duality gap The duality gap is the difference between the values of any primal solutions and any dual solutions If d displaystyle d nbsp is the optimal dual value and p displaystyle p nbsp is the optimal primal value then the duality gap is equal to p d displaystyle p d nbsp This value is always greater than or equal to 0 for minimization problems The duality gap is zero if and only if strong duality holds Otherwise the gap is strictly positive and weak duality holds 5 In computational optimization another duality gap is often reported which is the difference in value between any dual solution and the value of a feasible but suboptimal iterate for the primal problem This alternative duality gap quantifies the discrepancy between the value of a current feasible but suboptimal iterate for the primal problem and the value of the dual problem the value of the dual problem is under regularity conditions equal to the value of the convex relaxation of the primal problem The convex relaxation is the problem arising replacing a non convex feasible set with its closed convex hull and with replacing a non convex function with its convex closure that is the function that has the epigraph that is the closed convex hull of the original primal objective function 6 7 8 9 10 11 12 13 14 15 16 Linear case editMain article Dual linear program Linear programming problems are optimization problems in which the objective function and the constraints are all linear In the primal problem the objective function is a linear combination of n variables There are m constraints each of which places an upper bound on a linear combination of the n variables The goal is to maximize the value of the objective function subject to the constraints A solution is a vector a list of n values that achieves the maximum value for the objective function In the dual problem the objective function is a linear combination of the m values that are the limits in the m constraints from the primal problem There are n dual constraints each of which places a lower bound on a linear combination of m dual variables Relationship between the primal problem and the dual problem edit In the linear case in the primal problem from each sub optimal point that satisfies all the constraints there is a direction or subspace of directions to move that increases the objective function Moving in any such direction is said to remove slack between the candidate solution and one or more constraints An infeasible value of the candidate solution is one that exceeds one or more of the constraints In the dual problem the dual vector multiplies the constraints that determine the positions of the constraints in the primal Varying the dual vector in the dual problem is equivalent to revising the upper bounds in the primal problem The lowest upper bound is sought That is the dual vector is minimized in order to remove slack between the candidate positions of the constraints and the actual optimum An infeasible value of the dual vector is one that is too low It sets the candidate positions of one or more of the constraints in a position that excludes the actual optimum This intuition is made formal by the equations in Linear programming Duality Nonlinear case editIn nonlinear programming the constraints are not necessarily linear Nonetheless many of the same principles apply To ensure that the global maximum of a non linear problem can be identified easily the problem formulation often requires that the functions be convex and have compact lower level sets This is the significance of the Karush Kuhn Tucker conditions They provide necessary conditions for identifying local optima of non linear programming problems There are additional conditions constraint qualifications that are necessary so that it will be possible to define the direction to an optimal solution An optimal solution is one that is a local optimum but possibly not a global optimum Lagrange duality edit Motivation 17 Suppose we want to solve the following nonlinear programming problem minimize f 0 x subject to f i x 0 i 1 m displaystyle begin aligned text minimize amp f 0 x text subject to amp f i x leq 0 i in left 1 ldots m right end aligned nbsp The problem has constraints we would like to convert it to a program without constraints Theoretically it is possible to do it by minimizing the function J x defined asJ x f 0 x i I f i x displaystyle J x f 0 x sum i I f i x nbsp where I is an infinite step function I u 0 if u 0 and I u otherwise But J x is hard to solve as it is not continuous It is possible to approximate I u by lu where l is a positive constant This yields a function known as the lagrangian L x l f 0 x i l i f i x displaystyle L x lambda f 0 x sum i lambda i f i x nbsp Note that for every x max l 0 L x l J x displaystyle max lambda geq 0 L x lambda J x nbsp Proof If x satisfies all constraints fi x 0 then L x l is maximized when taking l 0 and its value is then f x If x violates some constraint fi x gt 0 for some i then L x l when li Therefore the original problem is equivalent to min x max l 0 L x l displaystyle min x max lambda geq 0 L x lambda nbsp By reversing the order of min and max we get max l 0 min x L x l displaystyle max lambda geq 0 min x L x lambda nbsp The dual function is the inner problem in the above formula g l min x L x l displaystyle g lambda min x L x lambda nbsp The Lagrangian dual program is the program of maximizing g max l 0 g l displaystyle max lambda geq 0 g lambda nbsp The optimal solution to the dual program is a lower bound for the optimal solution of the original primal program this is the weak duality principle If the primal problem is convex and bounded from below and there exists a point in which all nonlinear constraints are strictly satisfied Slater s condition then the optimal solution to the dual program equals the optimal solution of the primal program this is the strong duality principle In this case we can solve the primal program by finding an optimal solution l to the dual program and then solving min x L x l displaystyle min x L x lambda nbsp Note that to use either the weak or the strong duality principle we need a way to compute g l In general this may be hard as we need to solve a different minimization problem for every l But for some classes of functions it is possible to get an explicit formula for g Solving the primal and dual programs together is often easier than solving only one of them Examples are linear programming and quadratic programming A better and more general approach to duality is provided by Fenchel s duality theorem 18 Sub 3 3 1 Another condition in which the min max and max min are equal is when the Lagrangian has a saddle point x l is a saddle point of the Lagrange function L if and only if x is an optimal solution to the primal l is an optimal solution to the dual and the optimal values in the indicated problems are equal to each other 18 Prop 3 2 2 The strong Lagrange principle edit Given a nonlinear programming problem in standard form minimize f 0 x subject to f i x 0 i 1 m h i x 0 i 1 p displaystyle begin aligned text minimize amp f 0 x text subject to amp f i x leq 0 i in left 1 ldots m right amp h i x 0 i in left 1 ldots p right end aligned nbsp with the domain D R n displaystyle mathcal D subset mathbb R n nbsp having non empty interior the Lagrangian function L R n R m R p R displaystyle mathcal L mathbb R n times mathbb R m times mathbb R p to mathbb R nbsp is defined as L x l n f 0 x i 1 m l i f i x i 1 p n i h i x displaystyle mathcal L x lambda nu f 0 x sum i 1 m lambda i f i x sum i 1 p nu i h i x nbsp The vectors l displaystyle lambda nbsp and n displaystyle nu nbsp are called the dual variables or Lagrange multiplier vectors associated with the problem The Lagrange dual function g R m R p R displaystyle g mathbb R m times mathbb R p to mathbb R nbsp is defined as g l n inf x D L x l n inf x D f 0 x i 1 m l i f i x i 1 p n i h i x displaystyle g lambda nu inf x in mathcal D mathcal L x lambda nu inf x in mathcal D left f 0 x sum i 1 m lambda i f i x sum i 1 p nu i h i x right nbsp The dual function g is concave even when the initial problem is not convex because it is a point wise infimum of affine functions The dual function yields lower bounds on the optimal value p displaystyle p nbsp of the initial problem for any l 0 displaystyle lambda geq 0 nbsp and any n displaystyle nu nbsp we have g l n p displaystyle g lambda nu leq p nbsp If a constraint qualification such as Slater s condition holds and the original problem is convex then we have strong duality i e d max l 0 n g l n inf f 0 p displaystyle d max lambda geq 0 nu g lambda nu inf f 0 p nbsp Convex problems edit For a convex minimization problem with inequality constraints minimize x f x s u b j e c t t o g i x 0 i 1 m displaystyle begin aligned amp underset x operatorname minimize amp amp f x amp operatorname subject to amp amp g i x leq 0 quad i 1 ldots m end aligned nbsp the Lagrangian dual problem is maximize u inf x f x j 1 m u j g j x s u b j e c t t o u i 0 i 1 m displaystyle begin aligned amp underset u operatorname maximize amp amp inf x left f x sum j 1 m u j g j x right amp operatorname subject to amp amp u i geq 0 quad i 1 ldots m end aligned nbsp where the objective function is the Lagrange dual function Provided that the functions f displaystyle f nbsp and g 1 g m displaystyle g 1 ldots g m nbsp are continuously differentiable the infimum occurs where the gradient is equal to zero The problem maximize x u f x j 1 m u j g j x s u b j e c t t o f x j 1 m u j g j x 0 u i 0 i 1 m displaystyle begin aligned amp underset x u operatorname maximize amp amp f x sum j 1 m u j g j x amp operatorname subject to amp amp nabla f x sum j 1 m u j nabla g j x 0 amp amp amp u i geq 0 quad i 1 ldots m end aligned nbsp is called the Wolfe dual problem This problem may be difficult to deal with computationally because the objective function is not concave in the joint variables u x displaystyle u x nbsp Also the equality constraint f x j 1 m u j g j x displaystyle nabla f x sum j 1 m u j nabla g j x nbsp is nonlinear in general so the Wolfe dual problem is typically a nonconvex optimization problem In any case weak duality holds 19 History editAccording to George Dantzig the duality theorem for linear optimization was conjectured by John von Neumann immediately after Dantzig presented the linear programming problem Von Neumann noted that he was using information from his game theory and conjectured that two person zero sum matrix game was equivalent to linear programming Rigorous proofs were first published in 1948 by Albert W Tucker and his group Dantzig s foreword to Nering and Tucker 1993 Applications editIn support vector machines SVMs the formulating the primal problem of SVMs as the dual problem can be used to implement Kernel trick but the latter has higher time complexity in the historical cases See also editConvex duality Duality Relaxation approximation Notes edit Boyd Stephen P Vandenberghe Lieven 2004 Convex Optimization pdf Cambridge University Press p 216 ISBN 978 0 521 83378 3 Retrieved October 15 2011 a b Boţ Radu Ioan Wanka Gert Grad Sorin Mihai 2009 Duality in Vector Optimization Springer ISBN 978 3 642 02885 4 Csetnek Erno Robert 2010 Overcoming the failure of the classical generalized interior point regularity conditions in convex optimization Applications of the duality theory to enlargements of maximal monotone operators Logos Verlag Berlin GmbH ISBN 978 3 8325 2503 3 Zălinescu Constantin 2002 Convex analysis in general vector spaces River Edge NJ World Scientific Publishing Co Inc pp 106 113 ISBN 981 238 067 1 MR 1921556 Borwein Jonathan Zhu Qiji 2005 Techniques of Variational Analysis Springer ISBN 978 1 4419 2026 3 Ahuja Ravindra K Magnanti Thomas L Orlin James B 1993 Network Flows Theory Algorithms and Applications Prentice Hall ISBN 0 13 617549 X Bertsekas Dimitri Nedic Angelia Ozdaglar Asuman 2003 Convex Analysis and Optimization Athena Scientific ISBN 1 886529 45 0 Bertsekas Dimitri P 1999 Nonlinear Programming 2nd ed Athena Scientific ISBN 1 886529 00 0 Bertsekas Dimitri P 2009 Convex Optimization Theory Athena Scientific ISBN 978 1 886529 31 1 Bonnans J Frederic Gilbert J Charles Lemarechal Claude Sagastizabal Claudia A 2006 Numerical optimization Theoretical and practical aspects Universitext Second revised ed of translation of 1997 French ed Berlin Springer Verlag pp xiv 490 doi 10 1007 978 3 540 35447 5 ISBN 3 540 35445 X MR 2265882 Hiriart Urruty Jean Baptiste Lemarechal Claude 1993 Convex analysis and minimization algorithms Volume I Fundamentals Grundlehren der Mathematischen Wissenschaften Fundamental Principles of Mathematical Sciences Vol 305 Berlin Springer Verlag pp xviii 417 ISBN 3 540 56850 6 MR 1261420 Hiriart Urruty Jean Baptiste Lemarechal Claude 1993 14 Duality for Practitioners Convex analysis and minimization algorithms Volume II Advanced theory and bundle methods Grundlehren der Mathematischen Wissenschaften Fundamental Principles of Mathematical Sciences Vol 306 Berlin Springer Verlag pp xviii 346 ISBN 3 540 56852 2 MR 1295240 Lasdon Leon S 2002 Reprint of the 1970 Macmillan Optimization theory for large systems Mineola New York Dover Publications Inc pp xiii 523 ISBN 978 0 486 41999 2 MR 1888251 Lemarechal Claude 2001 Lagrangian relaxation In Junger Michael Naddef Denis eds Computational combinatorial optimization Papers from the Spring School held in Schloss Dagstuhl May 15 19 2000 Lecture Notes in Computer Science LNCS Vol 2241 Berlin Springer Verlag pp 112 156 doi 10 1007 3 540 45586 8 4 ISBN 3 540 42877 1 MR 1900016 S2CID 9048698 Minoux Michel 1986 Mathematical programming Theory and algorithms Egon Balas forward Steven Vajda trans from the 1983 Paris Dunod French Chichester A Wiley Interscience Publication John Wiley amp Sons Ltd pp xxviii 489 ISBN 0 471 90170 9 MR 0868279 2008 Second ed in French Programmation mathematique Theorie et algorithmes Editions Tec amp Doc Paris 2008 xxx 711 pp Shapiro Jeremy F 1979 Mathematical programming Structures and algorithms New York Wiley Interscience John Wiley amp Sons pp xvi 388 ISBN 0 471 77886 9 MR 0544669 David Knowles 2010 Lagrangian Duality for Dummies PDF a b Nemirovsky and Ben Tal 2023 Optimization III Convex Optimization PDF Geoffrion Arthur M 1971 Duality in Nonlinear Programming A Simplified Applications Oriented Development SIAM Review 13 1 1 37 doi 10 1137 1013001 JSTOR 2028848 References editBooks edit Ahuja Ravindra K Magnanti Thomas L Orlin James B 1993 Network Flows Theory Algorithms and Applications Prentice Hall ISBN 0 13 617549 X Bertsekas Dimitri Nedic Angelia Ozdaglar Asuman 2003 Convex Analysis and Optimization Athena Scientific ISBN 1 886529 45 0 Bertsekas Dimitri P 1999 Nonlinear Programming 2nd ed Athena Scientific ISBN 1 886529 00 0 Bertsekas Dimitri P 2009 Convex Optimization Theory Athena Scientific ISBN 978 1 886529 31 1 Bonnans J Frederic Gilbert J Charles Lemarechal Claude Sagastizabal Claudia A 2006 Numerical optimization Theoretical and practical aspects Universitext Second revised ed of translation of 1997 French ed Berlin Springer Verlag pp xiv 490 doi 10 1007 978 3 540 35447 5 ISBN 3 540 35445 X MR 2265882 Cook William J Cunningham William H Pulleyblank William R Schrijver Alexander November 12 1997 Combinatorial Optimization 1st ed John Wiley amp Sons ISBN 0 471 55894 X Dantzig George B 1963 Linear Programming and Extensions Princeton NJ Princeton University Press Hiriart Urruty Jean Baptiste Lemarechal Claude 1993 Convex analysis and minimization algorithms Volume I Fundamentals Grundlehren der Mathematischen Wissenschaften Fundamental Principles of Mathematical Sciences Vol 305 Berlin Springer Verlag pp xviii 417 ISBN 3 540 56850 6 MR 1261420 Hiriart Urruty Jean Baptiste Lemarechal Claude 1993 14 Duality for Practitioners Convex analysis and minimization algorithms Volume II Advanced theory and bundle methods Grundlehren der Mathematischen Wissenschaften Fundamental Principles of Mathematical Sciences Vol 306 Berlin Springer Verlag pp xviii 346 ISBN 3 540 56852 2 MR 1295240 Lasdon Leon S 2002 Reprint of the 1970 Macmillan Optimization theory for large systems Mineola New York Dover Publications Inc pp xiii 523 ISBN 978 0 486 41999 2 MR 1888251 Lawler Eugene 2001 4 5 Combinatorial Implications of Max Flow Min Cut Theorem 4 6 Linear Programming Interpretation of Max Flow Min Cut Theorem Combinatorial Optimization Networks and Matroids Dover pp 117 120 ISBN 0 486 41453 1 Lemarechal Claude 2001 Lagrangian relaxation In Junger Michael Naddef Denis eds Computational combinatorial optimization Papers from the Spring School held in Schloss Dagstuhl May 15 19 2000 Lecture Notes in Computer Science LNCS Vol 2241 Berlin Springer Verlag pp 112 156 doi 10 1007 3 540 45586 8 4 ISBN 3 540 42877 1 MR 1900016 S2CID 9048698 Minoux Michel 1986 Mathematical programming Theory and algorithms Egon Balas forward Steven Vajda trans from the 1983 Paris Dunod French Chichester A Wiley Interscience Publication John Wiley amp Sons Ltd pp xxviii 489 ISBN 0 471 90170 9 MR 0868279 2008 Second ed in French Programmation mathematique Theorie et algorithmes Editions Tec amp Doc Paris 2008 xxx 711 pp Nering Evar D Tucker Albert W 1993 Linear Programming and Related Problems Boston MA Academic Press ISBN 978 0 12 515440 6 Papadimitriou Christos H Steiglitz Kenneth July 1998 Combinatorial Optimization Algorithms and Complexity Unabridged ed Dover ISBN 0 486 40258 4 Ruszczynski Andrzej 2006 Nonlinear Optimization Princeton NJ Princeton University Press pp xii 454 ISBN 978 0 691 11915 1 MR 2199043 Articles edit Everett Hugh III 1963 Generalized Lagrange multiplier method for solving problems of optimum allocation of resources Operations Research 11 3 399 417 doi 10 1287 opre 11 3 399 JSTOR 168028 MR 0152360 Archived from the original on 2011 07 24 Kiwiel Krzysztof C Larsson Torbjorn Lindberg P O August 2007 Lagrangian relaxation via ballstep subgradient methods Mathematics of Operations Research 32 3 669 686 doi 10 1287 moor 1070 0261 MR 2348241 Archived from the original on 2011 07 26 Retrieved 2011 05 12 Duality in Linear Programming Gary D Knott Retrieved from https en wikipedia org w index php title Duality optimization amp oldid 1218987895, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.