fbpx
Wikipedia

Ellipsoid method

In mathematical optimization, the ellipsoid method is an iterative method for minimizing convex functions over convex sets. The ellipsoid method generates a sequence of ellipsoids whose volume uniformly decreases at every step, thus enclosing a minimizer of a convex function.

When specialized to solving feasible linear optimization problems with rational data, the ellipsoid method is an algorithm which finds an optimal solution in a number of steps that is polynomial in the input size.

History edit

The ellipsoid method has a long history. As an iterative method, a preliminary version was introduced by Naum Z. Shor. In 1972, an approximation algorithm for real convex minimization was studied by Arkadi Nemirovski and David B. Yudin (Judin).

As an algorithm for solving linear programming problems with rational data, the ellipsoid algorithm was studied by Leonid Khachiyan; Khachiyan's achievement was to prove the polynomial-time solvability of linear programs. This was a notable step from a theoretical perspective: The standard algorithm for solving linear problems at the time was the simplex algorithm, which has a run time that typically is linear in the size of the problem, but for which examples exist for which it is exponential in the size of the problem. As such, having an algorithm that is guaranteed to be polynomial for all cases seemed like a theoretical breakthrough.

Khachiyan's work showed, for the first time, that there can be algorithms for solving linear programs whose runtime can be proven to be polynomial. In practice, however, the algorithm is fairly slow and of little practical interest, though it provided inspiration for later work that turned out to be of much greater practical use. Specifically, Karmarkar's algorithm, an interior-point method, is much faster than the ellipsoid method in practice. Karmarkar's algorithm is also faster in the worst case.

The ellipsoidal algorithm allows complexity theorists to achieve (worst-case) bounds that depend on the dimension of the problem and on the size of the data, but not on the number of rows, so it remained important in combinatorial optimization theory for many years.[1][2][3][4] Only in the 21st century have interior-point algorithms with similar complexity properties appeared.[citation needed]

Description edit

A convex minimization problem consists of the following ingredients.

  • A convex function   to be minimized over the vector   (containing n variables);
  • Convex inequality constraints of the form  , where the functions   are convex; these constraints define a convex set  .
  • Linear equality constraints of the form  .

We are also given an initial ellipsoid   defined as

 

containing a minimizer  , where   and   is the center of  .

Finally, we require the existence of a separation oracle for the convex set  . Given a point  , the oracle should return one of two answers:[5]

  • "The point   is in  ", or -
  • "The point   is not in  , and moreover, here is a hyperplane that separates   from  ", that is, a vector   such that   for all  .

The output of the ellipsoid method is either:

  • Any point in the polytope   (i.e., any feasible point), or -
  • A proof that   is empty.

Inequality-constrained minimization of a function that is zero everywhere corresponds to the problem of simply identifying any feasible point. It turns out that any linear programming problem can be reduced to a linear feasibility problem (e.g. minimize the zero function subject to some linear inequality and equality constraints). One way to do this is by combining the primal and dual linear programs together into one program, and adding the additional (linear) constraint that the value of the primal solution is no worse than the value of the dual solution. Another way is to treat the objective of the linear program as an additional constraint, and use binary search to find the optimum value.[citation needed]

Unconstrained minimization edit

At the k-th iteration of the algorithm, we have a point   at the center of an ellipsoid

 

We query the cutting-plane oracle to obtain a vector   such that

 

We therefore conclude that

 

We set   to be the ellipsoid of minimal volume containing the half-ellipsoid described above and compute  . The update is given by

 

where

 

The stopping criterion is given by the property that

 
Sample sequence of iterates

Inequality-constrained minimization edit

At the k-th iteration of the algorithm for constrained minimization, we have a point   at the center of an ellipsoid   as before. We also must maintain a list of values   recording the smallest objective value of feasible iterates so far. Depending on whether or not the point   is feasible, we perform one of two tasks:

  • If   is feasible, perform essentially the same update as in the unconstrained case, by choosing a subgradient   that satisfies
 
  • If   is infeasible and violates the j-th constraint, update the ellipsoid with a feasibility cut. Our feasibility cut may be a subgradient   of   which must satisfy
 

for all feasible z.

Performance in convex programs edit

Theoretical run-time complexity guarantee edit

The run-time complexity guarantee of the ellipsoid method in the real RAM model is given by the following theorem.[6]: Thm.8.3.1 

Consider a family of convex optimization problems of the form: minimize f(x) s.t. x is in G, where f is a convex function and G is a convex set (a subset of an Euclidean space Rn). Each problem p in the family is represented by a data-vector Data(p), e.g., the real-valued coefficients in matrices and vectors representing the function f and the feasible region G. The size of a problem p, Size(p), is defined as the number of elements (real numbers) in Data(p). The following assumptions are needed:

  1. G (the feasible region) is:
    • Bounded;
    • Has a non-empty interior (so there is a strictly-feasible point);
  2. Given Data(p), one can compute using poly(Size(p)) arithmetic operations:
    • An ellipsoid that contains G;
    • A lower bound MinVol(p)>0 on the volume of G.
  3. Given Data(p) and a point x in Rn, one can compute using poly(Size(p)) arithmetic operations:
    • A separation oracle for G (that is: either assert that x is in G, or return a hyperplane separating x from G).
    • A first-order oracle for f (that is: compute the value of f(x) and a subgradient f'(x)).

Under these assumptions, the ellipsoid method is "R-polynomial". This means that there exists a polynomial Poly such that, for every problem-instance p and every approximation-ratio ε>0, the method finds a solution x satisfying :

 ,

using at most the following number of arithmetic operations on real numbers:

 

where V(p) is a data-dependent quantity. Intuitively, it means that the number of operations required for each additional digit of accuracy is polynomial in Size(p). In the case of the ellipsoid method, we have:

 .

The ellipsoid method requires at most   steps, and each step requires Poly(Size(p)) arithmetic operations.

Practical performance edit

The ellipsoid method is used on low-dimensional problems, such as planar location problems, where it is numerically stable. Nemirovsky and BenTal[6]: Sec.8.3.3  say that it is efficient if the number of variables is at most 20-30; this is so even if there are thousands of constraints, as the number of iterations does not depend on the number of constraints. However, in problems with many variables, the ellipsoid method is very inefficient, as the number of iterations grows as O(n2).

Even on "small"-sized problems, it suffers from numerical instability and poor performance in practice[citation needed].

Theoretical importance edit

The ellipsoid method is an important theoretical technique in combinatorial optimization. In computational complexity theory, the ellipsoid algorithm is attractive because its complexity depends on the number of columns and the digital size of the coefficients, but not on the number of rows.

The ellipsoid method can be used to show that many algorithmic problems on convex sets are polynomial-time equivalent.

Performance in linear programs edit

Leonid Khachiyan applied the ellipsoid method to the special case of linear programming: minimize cTx s.t. Ax ≤ b, where all coefficients in A,b,c are rational numbers. He showed that linear programs can be solved in polynomial time. Here is a sketch of Khachiyan's theorem.[6]: Sec.8.4.2 

Step 1: reducing optimization to search. The theorem of linear programming duality says that we can reduce the above minimization problem to the search problem: find x,y s.t. Ax ≤ b ; ATy = c ; y ≤ 0 ; cTx=bTy. The first problem is solvable iff the second problem is solvable; in case the problem is solvable, the x-components of the solution to the second problem are an optimal solution to the first problem. Therefore, from now on, we can assume that we need to solve the following problem: find z ≥ 0 s.t. Rzr. Multiplying all rational coefficients by the common denominator, we can assume that all coefficients are integers.

Step 2: reducing search to feasibility-check. The problem find z ≥ 0 s.t. Rzr can be reduced to the binary decision problem: "is there a z ≥ 0 such that Rzr?". This can be done as follows. If the answer to the decision problem is "no", then the answer to the search problem is "None", and we are done. Otherwise, take the first inequality constraint R1zr1; replace it with an equality R1z = r1; and apply the decision problem again. If the answer is "yes", we keep the equality; if the answer is "no", it means that the inequality is redundant, and we can remove it. Then we proceed to the next inequality constraint. For each constraint, we either convert it to equality or remove it. Finally, we have only equality constraints, which can be solved by any method for solving a system of linear equations.

Step 3: the decision problem can be reduced to a different optimization problem. Define the residual function f(z) := max[(Rz)1-r1, (Rz)2-r2, (Rz)3-r3,...]. Clearly, f(z)≤0 iff Rzr. Therefore, to solve the decnsion problem, it is sufficient to solve the minimization problem: minz f(z). The function f is convex (it is a maximum of linear functions). Denote the minimum value by f*. Then the answer to the decision problem is "yes" iff f*≤0.

Step 4: In the optimization problem minz f(z), we can assume that z is in a box of side-length 2L, where L is the bit length of the problem data. Thus, we have a bounded convex program, that can be solved up to any accuracy ε by the ellipsoid method, in time polynomial in L.

Step 5: It can be proved that, if f*>0, then f*>2-poly(L), for some polynomial. Therefore, we can pick the accuracy ε=2-poly(L). Then, the ε-approximate solution found by the ellipsoid method will be positive, iff f*>0, iff the decision problem is unsolvable.

Variants edit

The ellipsoid method has several variants, depending on what cuts exactly are used in each step.[1]: Sec. 3 

Different cuts edit

In the central-cut ellipsoid method,[1]: 82, 87–94  the cuts are always through the center of the current ellipsoid. The input is a rational number ε>0, a convex body K given by a weak separation oracle, and a number R such that S(0,R) (the ball of radius R around the origin) contains K. The output is one of the following:

  • (a) A vector at a distance of at most ε from K, or --
  • (b) A positive definite matrix A and a point a such that the ellipsoid E(A,a) contains K, and the volume of E(A,a) is at most ε.

The number of steps is  , the number of required accuracy digits is p := 8N, and the required accuracy of the separation oracle is d := 2-p.

In the deep-cut ellipsoid method,[1]: 83  the cuts remove more than half of the ellipsoid in each step. This makes it faster to discover that K is empty. However, when K is nonempty, there are examples in which the central-cut method finds a feasible point faster. The use of deep cuts does not change the order of magnitude of the run-time.

In the shallow-cut ellipsoid method,[1]: 83, 94–101  the cuts remove less than half of the ellipsoid in each step. This variant is not very useful in practice, but it has theoretical importance: it allows to prove results that cannot be derived from other variants. The input is a rational number ε>0, a convex body K given by a shallow separation oracle, and a number R such that S(0,R) contains K. The output is a positive definite matrix A and a point a such that one of the following holds:

  • (a) The ellipsoid E(A,a) has been declared "tough" by the oracle, or -
  • (b) K is contained in E(A,a) and the volume of E(A,a) is at most ε.

The number of steps is  , and the number of required accuracy digits is p := 8N.

Different ellipsoids edit

There is also a distinction between the circumscribed ellipsoid and the inscribed ellipsoid methods:[7]

  • In the circumscribed ellipsoid method, each iteration finds an ellipsoid of smallest volume that contains the remaining part of the previous ellipsoid. This method was developed by Yudin and Nemirovskii.[8]
  • In the Inscribed ellipsoid method, each iteration finds an ellipsoid of largest volume that is contained the remaining part of the previous ellipsoid. This method was developed by Tarasov, Khachian and Erlikh.[9]

The methods differ in their runtime complexity (below, n is the number of variables and epsilon is the accuracy):

  • The circumscribed method requires   iterations, where each iteration consists of finding a separating hyperplane and finding a new circumscribed ellipsoid. Finding a circumscribed ellipsoid requires   time.
  • The inscribed method requires   iterations, where each iteration consists of finding a separating hyperplane and finding a new inscribed ellipsoid. Finding an inscribed ellipsoid requires   time for some small  .

The relative efficiency of the methods depends on the time required for finding a separating hyperplane, which depends on the application: if the runtime is   for   then the circumscribed method is more efficient, but if   then the inscribed method is more efficient.[7]

Related methods edit

  • The center-of-gravity method is a conceptually simpler method, that requires fewer steps. However, each step is computationally expensive, as it requires to compute the center of gravity of the current feasible polytope.
  • Interior point methods, too, allow solving convex optimization problems in polynomial time, but their practical performance is much better than the ellipsoid method.

Notes edit

  1. ^ a b c d e Grötschel, Martin; Lovász, László; Schrijver, Alexander (1993), Geometric algorithms and combinatorial optimization, Algorithms and Combinatorics, vol. 2 (2nd ed.), Springer-Verlag, Berlin, doi:10.1007/978-3-642-78240-4, ISBN 978-3-642-78242-8, MR 1261419
  2. ^ L. Lovász: An Algorithmic Theory of Numbers, Graphs, and Convexity, CBMS-NSF Regional Conference Series in Applied Mathematics 50, SIAM, Philadelphia, Pennsylvania, 1986.
  3. ^ V. Chandru and M.R.Rao, Linear Programming, Chapter 31 in Algorithms and Theory of Computation Handbook, edited by M. J. Atallah, CRC Press 1999, 31-1 to 31-37.
  4. ^ V. Chandru and M.R.Rao, Integer Programming, Chapter 32 in Algorithms and Theory of Computation Handbook, edited by M.J.Atallah, CRC Press 1999, 32-1 to 32-45.
  5. ^ "MIT 6.854 Spring 2016 Lecture 12: From Separation to Optimization and Back; Ellipsoid Method - YouTube". www.youtube.com. Archived from the original on 2021-12-22. Retrieved 2021-01-03.
  6. ^ a b c Nemirovsky and Ben-Tal (2023). "Optimization III: Convex Optimization" (PDF).[permanent dead link]
  7. ^ a b Newman, D. J.; Primak, M. E. (1992-12-01). "Complexity of circumscribed and inscribed ellipsoid methods for solving equilibrium economical models". Applied Mathematics and Computation. 52 (2): 223–231. doi:10.1016/0096-3003(92)90079-G. ISSN 0096-3003.
  8. ^ https://elibrary.ru/item.asp?id=38308898
  9. ^ Primak, M. E.; Kheyfets, B. L. (1995-06-01). "A modification of the inscribed ellipsoid method". Mathematical and Computer Modelling. 21 (11): 69–76. doi:10.1016/0895-7177(95)00080-L. ISSN 0895-7177.

Further reading edit

  • Dmitris Alevras and Manfred W. Padberg, Linear Optimization and Extensions: Problems and Extensions, Universitext, Springer-Verlag, 2001. (Problems from Padberg with solutions.)
  • V. Chandru and M.R.Rao, Linear Programming, Chapter 31 in Algorithms and Theory of Computation Handbook, edited by M.J.Atallah, CRC Press 1999, 31-1 to 31-37.
  • V. Chandru and M.R.Rao, Integer Programming, Chapter 32 in Algorithms and Theory of Computation Handbook, edited by M.J.Atallah, CRC Press 1999, 32-1 to 32-45.
  • George B. Dantzig and Mukund N. Thapa. 1997. Linear programming 1: Introduction. Springer-Verlag.
  • George B. Dantzig and Mukund N. Thapa. 2003. Linear Programming 2: Theory and Extensions. Springer-Verlag.
  • L. Lovász: An Algorithmic Theory of Numbers, Graphs, and Convexity, CBMS-NSF Regional Conference Series in Applied Mathematics 50, SIAM, Philadelphia, Pennsylvania, 1986
  • Kattta G. Murty, Linear Programming, Wiley, 1983.
  • M. Padberg, Linear Optimization and Extensions, Second Edition, Springer-Verlag, 1999.
  • Christos H. Papadimitriou and Kenneth Steiglitz, Combinatorial Optimization: Algorithms and Complexity, Corrected republication with a new preface, Dover.
  • Alexander Schrijver, Theory of Linear and Integer Programming. John Wiley & sons, 1998, ISBN 0-471-98232-6

External links edit

  • EE364b, a Stanford course homepage

ellipsoid, method, mathematical, optimization, ellipsoid, method, iterative, method, minimizing, convex, functions, over, convex, sets, ellipsoid, method, generates, sequence, ellipsoids, whose, volume, uniformly, decreases, every, step, thus, enclosing, minim. In mathematical optimization the ellipsoid method is an iterative method for minimizing convex functions over convex sets The ellipsoid method generates a sequence of ellipsoids whose volume uniformly decreases at every step thus enclosing a minimizer of a convex function When specialized to solving feasible linear optimization problems with rational data the ellipsoid method is an algorithm which finds an optimal solution in a number of steps that is polynomial in the input size Contents 1 History 2 Description 3 Unconstrained minimization 4 Inequality constrained minimization 5 Performance in convex programs 5 1 Theoretical run time complexity guarantee 5 2 Practical performance 5 3 Theoretical importance 6 Performance in linear programs 7 Variants 7 1 Different cuts 7 2 Different ellipsoids 8 Related methods 9 Notes 10 Further reading 11 External linksHistory editThe ellipsoid method has a long history As an iterative method a preliminary version was introduced by Naum Z Shor In 1972 an approximation algorithm for real convex minimization was studied by Arkadi Nemirovski and David B Yudin Judin As an algorithm for solving linear programming problems with rational data the ellipsoid algorithm was studied by Leonid Khachiyan Khachiyan s achievement was to prove the polynomial time solvability of linear programs This was a notable step from a theoretical perspective The standard algorithm for solving linear problems at the time was the simplex algorithm which has a run time that typically is linear in the size of the problem but for which examples exist for which it is exponential in the size of the problem As such having an algorithm that is guaranteed to be polynomial for all cases seemed like a theoretical breakthrough Khachiyan s work showed for the first time that there can be algorithms for solving linear programs whose runtime can be proven to be polynomial In practice however the algorithm is fairly slow and of little practical interest though it provided inspiration for later work that turned out to be of much greater practical use Specifically Karmarkar s algorithm an interior point method is much faster than the ellipsoid method in practice Karmarkar s algorithm is also faster in the worst case The ellipsoidal algorithm allows complexity theorists to achieve worst case bounds that depend on the dimension of the problem and on the size of the data but not on the number of rows so it remained important in combinatorial optimization theory for many years 1 2 3 4 Only in the 21st century have interior point algorithms with similar complexity properties appeared citation needed Description editMain article Convex optimization A convex minimization problem consists of the following ingredients A convex function f0 x Rn R displaystyle f 0 x mathbb R n to mathbb R nbsp to be minimized over the vector x displaystyle x nbsp containing n variables Convex inequality constraints of the form fi x 0 displaystyle f i x leqslant 0 nbsp where the functions fi displaystyle f i nbsp are convex these constraints define a convex set Q displaystyle Q nbsp Linear equality constraints of the form hi x 0 displaystyle h i x 0 nbsp We are also given an initial ellipsoid E 0 Rn displaystyle mathcal E 0 subset mathbb R n nbsp defined as E 0 z Rn z x0 TP 0 1 z x0 1 displaystyle mathcal E 0 left z in mathbb R n z x 0 T P 0 1 z x 0 leqslant 1 right nbsp containing a minimizer x displaystyle x nbsp where P 0 0 displaystyle P 0 succ 0 nbsp and x0 displaystyle x 0 nbsp is the center of E displaystyle mathcal E nbsp Finally we require the existence of a separation oracle for the convex set Q displaystyle Q nbsp Given a point x Rn displaystyle x in mathbb R n nbsp the oracle should return one of two answers 5 The point x displaystyle x nbsp is in Q displaystyle Q nbsp or The point x displaystyle x nbsp is not in Q displaystyle Q nbsp and moreover here is a hyperplane that separates x displaystyle x nbsp from Q displaystyle Q nbsp that is a vector c displaystyle c nbsp such that c x lt c y displaystyle c cdot x lt c cdot y nbsp for all y Q displaystyle y in Q nbsp The output of the ellipsoid method is either Any point in the polytope Q displaystyle Q nbsp i e any feasible point or A proof that Q displaystyle Q nbsp is empty Inequality constrained minimization of a function that is zero everywhere corresponds to the problem of simply identifying any feasible point It turns out that any linear programming problem can be reduced to a linear feasibility problem e g minimize the zero function subject to some linear inequality and equality constraints One way to do this is by combining the primal and dual linear programs together into one program and adding the additional linear constraint that the value of the primal solution is no worse than the value of the dual solution Another way is to treat the objective of the linear program as an additional constraint and use binary search to find the optimum value citation needed Unconstrained minimization editAt the k th iteration of the algorithm we have a point x k displaystyle x k nbsp at the center of an ellipsoid E k x Rn x x k TP k 1 x x k 1 displaystyle mathcal E k left x in mathbb R n left x x k right T P k 1 left x x k right leqslant 1 right nbsp We query the cutting plane oracle to obtain a vector g k 1 Rn displaystyle g k 1 in mathbb R n nbsp such that g k 1 T x x k 0 displaystyle g k 1 T left x x k right leqslant 0 nbsp We therefore conclude that x E k z g k 1 T z x k 0 displaystyle x in mathcal E k cap left z g k 1 T left z x k right leqslant 0 right nbsp We set E k 1 displaystyle mathcal E k 1 nbsp to be the ellipsoid of minimal volume containing the half ellipsoid described above and compute x k 1 displaystyle x k 1 nbsp The update is given by x k 1 x k 1n 1P k g k 1 P k 1 n2n2 1 P k 2n 1P k g k 1 g k 1 TP k displaystyle begin aligned x k 1 amp x k frac 1 n 1 P k tilde g k 1 P k 1 amp frac n 2 n 2 1 left P k frac 2 n 1 P k tilde g k 1 tilde g k 1 T P k right end aligned nbsp where g k 1 1g k 1 TP k g k 1 g k 1 displaystyle tilde g k 1 left frac 1 sqrt g k 1 T P k g k 1 right g k 1 nbsp The stopping criterion is given by the property that g k TP k g k ϵ f x k f x ϵ displaystyle sqrt g k T P k g k leqslant epsilon quad Rightarrow quad f x k f left x right leqslant epsilon nbsp Sample sequence of iteratesInequality constrained minimization editAt the k th iteration of the algorithm for constrained minimization we have a point x k displaystyle x k nbsp at the center of an ellipsoid E k displaystyle mathcal E k nbsp as before We also must maintain a list of values fbest k displaystyle f rm best k nbsp recording the smallest objective value of feasible iterates so far Depending on whether or not the point x k displaystyle x k nbsp is feasible we perform one of two tasks If x k displaystyle x k nbsp is feasible perform essentially the same update as in the unconstrained case by choosing a subgradient g0 displaystyle g 0 nbsp that satisfiesg0T x x k f0 x k fbest k 0 displaystyle g 0 T x x k f 0 x k f rm best k leqslant 0 nbsp dd If x k displaystyle x k nbsp is infeasible and violates the j th constraint update the ellipsoid with a feasibility cut Our feasibility cut may be a subgradient gj displaystyle g j nbsp of fj displaystyle f j nbsp which must satisfygjT z x k fj x k 0 displaystyle g j T z x k f j x k leqslant 0 nbsp dd for all feasible z Performance in convex programs editTheoretical run time complexity guarantee edit The run time complexity guarantee of the ellipsoid method in the real RAM model is given by the following theorem 6 Thm 8 3 1 Consider a family of convex optimization problems of the form minimize f x s t x is in G where f is a convex function and G is a convex set a subset of an Euclidean space Rn Each problem p in the family is represented by a data vector Data p e g the real valued coefficients in matrices and vectors representing the function f and the feasible region G The size of a problem p Size p is defined as the number of elements real numbers in Data p The following assumptions are needed G the feasible region is Bounded Has a non empty interior so there is a strictly feasible point Given Data p one can compute using poly Size p arithmetic operations An ellipsoid that contains G A lower bound MinVol p gt 0 on the volume of G Given Data p and a point x in Rn one can compute using poly Size p arithmetic operations A separation oracle for G that is either assert that x is in G or return a hyperplane separating x from G A first order oracle for f that is compute the value of f x and a subgradient f x Under these assumptions the ellipsoid method is R polynomial This means that there exists a polynomial Poly such that for every problem instance p and every approximation ratio e gt 0 the method finds a solution x satisfying f x minGf e maxGf minGf displaystyle f x min G f leq varepsilon cdot max G f min G f nbsp using at most the following number of arithmetic operations on real numbers Poly Size p ln V p ϵ displaystyle Poly Size p cdot ln left frac V p epsilon right nbsp where V p is a data dependent quantity Intuitively it means that the number of operations required for each additional digit of accuracy is polynomial in Size p In the case of the ellipsoid method we have V p Vol initial ellipsoid Vol G 1 n Vol initial ellipsoid MinVol p 1 n displaystyle V p left frac Vol text initial ellipsoid Vol G right 1 n leq left frac Vol text initial ellipsoid MinVol p right 1 n nbsp The ellipsoid method requires at most 2 n 1 n ln V p ϵ displaystyle 2 n 1 n cdot ln left frac V p epsilon right nbsp steps and each step requires Poly Size p arithmetic operations Practical performance edit The ellipsoid method is used on low dimensional problems such as planar location problems where it is numerically stable Nemirovsky and BenTal 6 Sec 8 3 3 say that it is efficient if the number of variables is at most 20 30 this is so even if there are thousands of constraints as the number of iterations does not depend on the number of constraints However in problems with many variables the ellipsoid method is very inefficient as the number of iterations grows as O n2 Even on small sized problems it suffers from numerical instability and poor performance in practice citation needed Theoretical importance edit The ellipsoid method is an important theoretical technique in combinatorial optimization In computational complexity theory the ellipsoid algorithm is attractive because its complexity depends on the number of columns and the digital size of the coefficients but not on the number of rows The ellipsoid method can be used to show that many algorithmic problems on convex sets are polynomial time equivalent Performance in linear programs editLeonid Khachiyan applied the ellipsoid method to the special case of linear programming minimize cTx s t Ax b where all coefficients in A b c are rational numbers He showed that linear programs can be solved in polynomial time Here is a sketch of Khachiyan s theorem 6 Sec 8 4 2 Step 1 reducing optimization to search The theorem of linear programming duality says that we can reduce the above minimization problem to the search problem find x y s t Ax b ATy c y 0 cTx bTy The first problem is solvable iff the second problem is solvable in case the problem is solvable the x components of the solution to the second problem are an optimal solution to the first problem Therefore from now on we can assume that we need to solve the following problem find z 0 s t Rz r Multiplying all rational coefficients by the common denominator we can assume that all coefficients are integers Step 2 reducing search to feasibility check The problem find z 0 s t Rz r can be reduced to the binary decision problem is there a z 0 such that Rz r This can be done as follows If the answer to the decision problem is no then the answer to the search problem is None and we are done Otherwise take the first inequality constraint R1z r1 replace it with an equality R1z r1 and apply the decision problem again If the answer is yes we keep the equality if the answer is no it means that the inequality is redundant and we can remove it Then we proceed to the next inequality constraint For each constraint we either convert it to equality or remove it Finally we have only equality constraints which can be solved by any method for solving a system of linear equations Step 3 the decision problem can be reduced to a different optimization problem Define the residual function f z max Rz 1 r1 Rz 2 r2 Rz 3 r3 Clearly f z 0 iff Rz r Therefore to solve the decnsion problem it is sufficient to solve the minimization problem minz f z The function f is convex it is a maximum of linear functions Denote the minimum value by f Then the answer to the decision problem is yes iff f 0 Step 4 In the optimization problem minz f z we can assume that z is in a box of side length 2L where L is the bit length of the problem data Thus we have a bounded convex program that can be solved up to any accuracy e by the ellipsoid method in time polynomial in L Step 5 It can be proved that if f gt 0 then f gt 2 poly L for some polynomial Therefore we can pick the accuracy e 2 poly L Then the e approximate solution found by the ellipsoid method will be positive iff f gt 0 iff the decision problem is unsolvable Variants editThe ellipsoid method has several variants depending on what cuts exactly are used in each step 1 Sec 3 Different cuts edit In the central cut ellipsoid method 1 82 87 94 the cuts are always through the center of the current ellipsoid The input is a rational number e gt 0 a convex body K given by a weak separation oracle and a number R such that S 0 R the ball of radius R around the origin contains K The output is one of the following a A vector at a distance of at most e from K or b A positive definite matrix A and a point a such that the ellipsoid E A a contains K and the volume of E A a is at most e The number of steps is N 5nlog 1 ϵ 5n2log 2R displaystyle N lceil 5n log 1 epsilon 5n 2 log 2R rceil nbsp the number of required accuracy digits is p 8N and the required accuracy of the separation oracle is d 2 p In the deep cut ellipsoid method 1 83 the cuts remove more than half of the ellipsoid in each step This makes it faster to discover that K is empty However when K is nonempty there are examples in which the central cut method finds a feasible point faster The use of deep cuts does not change the order of magnitude of the run time In the shallow cut ellipsoid method 1 83 94 101 the cuts remove less than half of the ellipsoid in each step This variant is not very useful in practice but it has theoretical importance it allows to prove results that cannot be derived from other variants The input is a rational number e gt 0 a convex body K given by a shallow separation oracle and a number R such that S 0 R contains K The output is a positive definite matrix A and a point a such that one of the following holds a The ellipsoid E A a has been declared tough by the oracle or b K is contained in E A a and the volume of E A a is at most e The number of steps is N 5n n 1 2log 1 ϵ 5n2 n 1 2log 2R log n 1 displaystyle N lceil 5n n 1 2 log 1 epsilon 5n 2 n 1 2 log 2R log n 1 rceil nbsp and the number of required accuracy digits is p 8N Different ellipsoids edit There is also a distinction between the circumscribed ellipsoid and the inscribed ellipsoid methods 7 In the circumscribed ellipsoid method each iteration finds an ellipsoid of smallest volume that contains the remaining part of the previous ellipsoid This method was developed by Yudin and Nemirovskii 8 In the Inscribed ellipsoid method each iteration finds an ellipsoid of largest volume that is contained the remaining part of the previous ellipsoid This method was developed by Tarasov Khachian and Erlikh 9 The methods differ in their runtime complexity below n is the number of variables and epsilon is the accuracy The circumscribed method requires O n2 ln 1ϵ displaystyle O n 2 ln frac 1 epsilon nbsp iterations where each iteration consists of finding a separating hyperplane and finding a new circumscribed ellipsoid Finding a circumscribed ellipsoid requires O n2 displaystyle O n 2 nbsp time The inscribed method requires O n ln 1ϵ displaystyle O n ln frac 1 epsilon nbsp iterations where each iteration consists of finding a separating hyperplane and finding a new inscribed ellipsoid Finding an inscribed ellipsoid requires O n3 5 d displaystyle O n 3 5 delta nbsp time for some small d gt 0 displaystyle delta gt 0 nbsp The relative efficiency of the methods depends on the time required for finding a separating hyperplane which depends on the application if the runtime is O nt displaystyle O n t nbsp for t 2 5 displaystyle t leq 2 5 nbsp then the circumscribed method is more efficient but if t gt 2 5 displaystyle t gt 2 5 nbsp then the inscribed method is more efficient 7 Related methods editThe center of gravity method is a conceptually simpler method that requires fewer steps However each step is computationally expensive as it requires to compute the center of gravity of the current feasible polytope Interior point methods too allow solving convex optimization problems in polynomial time but their practical performance is much better than the ellipsoid method Notes edit a b c d e Grotschel Martin Lovasz Laszlo Schrijver Alexander 1993 Geometric algorithms and combinatorial optimization Algorithms and Combinatorics vol 2 2nd ed Springer Verlag Berlin doi 10 1007 978 3 642 78240 4 ISBN 978 3 642 78242 8 MR 1261419 L Lovasz An Algorithmic Theory of Numbers Graphs and Convexity CBMS NSF Regional Conference Series in Applied Mathematics 50 SIAM Philadelphia Pennsylvania 1986 V Chandru and M R Rao Linear Programming Chapter 31 in Algorithms and Theory of Computation Handbook edited by M J Atallah CRC Press 1999 31 1 to 31 37 V Chandru and M R Rao Integer Programming Chapter 32 in Algorithms and Theory of Computation Handbook edited by M J Atallah CRC Press 1999 32 1 to 32 45 MIT 6 854 Spring 2016 Lecture 12 From Separation to Optimization and Back Ellipsoid Method YouTube www youtube com Archived from the original on 2021 12 22 Retrieved 2021 01 03 a b c Nemirovsky and Ben Tal 2023 Optimization III Convex Optimization PDF permanent dead link a b Newman D J Primak M E 1992 12 01 Complexity of circumscribed and inscribed ellipsoid methods for solving equilibrium economical models Applied Mathematics and Computation 52 2 223 231 doi 10 1016 0096 3003 92 90079 G ISSN 0096 3003 https elibrary ru item asp id 38308898 Primak M E Kheyfets B L 1995 06 01 A modification of the inscribed ellipsoid method Mathematical and Computer Modelling 21 11 69 76 doi 10 1016 0895 7177 95 00080 L ISSN 0895 7177 Further reading editDmitris Alevras and Manfred W Padberg Linear Optimization and Extensions Problems and Extensions Universitext Springer Verlag 2001 Problems from Padberg with solutions V Chandru and M R Rao Linear Programming Chapter 31 in Algorithms and Theory of Computation Handbook edited by M J Atallah CRC Press 1999 31 1 to 31 37 V Chandru and M R Rao Integer Programming Chapter 32 in Algorithms and Theory of Computation Handbook edited by M J Atallah CRC Press 1999 32 1 to 32 45 George B Dantzig and Mukund N Thapa 1997 Linear programming 1 Introduction Springer Verlag George B Dantzig and Mukund N Thapa 2003 Linear Programming 2 Theory and Extensions Springer Verlag L Lovasz An Algorithmic Theory of Numbers Graphs and Convexity CBMS NSF Regional Conference Series in Applied Mathematics 50 SIAM Philadelphia Pennsylvania 1986 Kattta G Murty Linear Programming Wiley 1983 M Padberg Linear Optimization and Extensions Second Edition Springer Verlag 1999 Christos H Papadimitriou and Kenneth Steiglitz Combinatorial Optimization Algorithms and Complexity Corrected republication with a new preface Dover Alexander Schrijver Theory of Linear and Integer Programming John Wiley amp sons 1998 ISBN 0 471 98232 6External links editEE364b a Stanford course homepage Retrieved from https en wikipedia org w index php title Ellipsoid method amp oldid 1213669133, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.