fbpx
Wikipedia

Bregman method

The Bregman method is an iterative algorithm to solve certain convex optimization problems involving regularization.[1] The original version is due to Lev M. Bregman, who published it in 1967.[2]

The algorithm is a row-action method accessing constraint functions one by one and the method is particularly suited for large optimization problems where constraints can be efficiently enumerated[citation needed]. The algorithm works particularly well for regularizers such as the norm, where it converges very quickly because of an error cancellation effect.[3]

Algorithm

In order to be able to use the Bregman method, one must frame the problem of interest as finding  , where   is a regularizing function such as  .[3]

The Bregman distance is defined as   where   belongs to the subdifferential of   at   (which we denoted  ).[3][4] One performs the iteration  , with   a constant to be chosen by the user (and the minimization performed by an ordinary convex optimization algorithm),[3] or  , with   chosen each time to be a member of  .[4]

The algorithm starts with a pair of primal and dual variables. Then, for each constraint a generalized projection onto its feasible set is performed, updating both the constraint's dual variable and all primal variables for which there are non-zero coefficients in the constraint functions gradient. In case the objective is strictly convex and all constraint functions are convex, the limit of this iterative projection converges to the optimal primal dual pair.[citation needed]

In the case of a basis pursuit-type problem  , the Bregman method is equivalent to ordinary gradient descent on the dual problem  .[5] An exact regularization-type effect also occurs in this case; if   exceeds a certain threshold, the optimum value of   is precisely the optimum solution of  .[3][5]

Applications

The Bregman method or its generalizations can be applied to:

Generalizations and drawbacks

The method has links to the method of multipliers and dual ascent method (through the so-called Bregman alternating direction method of multipliers,[10][7] generalizing the alternating direction method of multipliers[8]) and multiple generalizations exist.

One drawback of the method is that it is only provably convergent if the objective function is strictly convex. In case this can not be ensured, as for linear programs or non-strictly convex quadratic programs, additional methods such as proximal gradient methods have been developed.[citation needed] In the case of the Rudin-Osher-Fatemi model of image denoising[clarification needed], the Bregman method provably converges.[11]

Some generalizations of the Bregman method include:

Linearized Bregman

In the Linearized Bregman method, one linearizes the intermediate objective functions   by replacing the second term with   (which approximates the second term near  ) and adding the penalty term   for a constant  . The result is much more computationally tractable, especially in basis pursuit-type problems.[4][5] In the case of a generic basis pursuit problem  , one can express the iteration as   for each component  , where we define  .[4]

Sometimes, when running the Linearized Bregman method, there are periods of "stagnation" where the residual[clarification needed] is almost constant. To alleviate this issue, one can use the Linearized Bregman method with kicking, where one essentially detects the beginning of a stagnation period, then predicts and skips to the end of it.[4][5]

Since Linearized Bregman is mathematically equivalent to gradient descent, it can be accelerated with methods to accelerate gradient descent, such as line search, L-BGFS, Barzilai-Borwein steps, or the Nesterov method; the last has been proposed as the accelerated linearized Bregman method.[5][9]

Split Bregman

The Split Bregman method solves problems of the form  , where   and   are both convex,[4] particularly problems of the form  .[6] We start by rewriting it as the constrained optimization problem  , then relax it into   where   is a constant. By defining  , one reduces the problem to one that can be solved with the ordinary Bregman algorithm.[4][6]

The Split Bregman method has been generalized to optimization over complex numbers using Wirtinger derivatives.[1]

References

  1. ^ a b c d Xiong, Kai; Zhao, Guanghui; Shi, Guangming; Wang, Yingbin (2019-09-12). "A Convex Optimization Algorithm for Compressed Sensing in a Complex Domain: The Complex-Valued Split Bregman Method". Sensors (published 18 Oct 2019). 19 (20): 4540. Bibcode:2019Senso..19.4540X. doi:10.3390/s19204540. PMC 6832202. PMID 31635423.
  2. ^ Bregman L. "A Relaxation Method of Finding a Common Point of Convex Sets and its Application to Problems of Optimization". Dokl. Akad. Nauk SSSR, v. 171, No. 5, 1966, p.p. 1019-1022. (English translation: Soviet Math. Dokl., v. 7, 1966, p.p. 1578-1581)
  3. ^ a b c d e f g h i j k Yin, Wotao (8 Dec 2009). "The Bregman Methods: Review and New Results" (PDF). (PDF) from the original on 2010-06-13. Retrieved 16 Apr 2021.
  4. ^ a b c d e f g h Bush, Jacqueline (10 Jun 2011). "University of California, Santa Barbara Senior Thesis: Bregman Algorithms" (PDF). University of California Santa Barbara. (PDF) from the original on 2016-11-30. Retrieved 16 Apr 2021.
  5. ^ a b c d e f Yin, Wotao (28 May 2009). "Analysis and Generalizations of the Linearized Bregman Method" (PDF). SIAM Journal on Imaging Sciences. 3 (4): 856–877. doi:10.1137/090760350. Retrieved 16 Apr 2021.
  6. ^ a b c Goldstein, Tom; Osher, Stanley (2 Jun 2008). "The Split Bregman Method for L1-Regularized Problems". SIAM J. Imaging Sci. 2 (2): 323–343. doi:10.1137/080725891. Retrieved 22 Apr 2021.
  7. ^ a b Jiang, Chunzhi (May 2015). "Comparison of Variable Penalty ADMM with Split Bregman Method on Hyperspectral Imaging Problems". from the original on 2020-03-23. Retrieved 20 Apr 2021.
  8. ^ a b c d Boyd, Stephen; Parikh, Neal; Chu, Eric; Peleato, Borja; Eckstein, Jonathan (19 Nov 2010). "Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers". Foundations and Trends in Machine Learning. 3: 1–122. CiteSeerX 10.1.1.722.981. doi:10.1561/2200000016.
  9. ^ a b Huang, Bo; Ma, Shiqian; Goldfarb, Donald (27 Jun 2011). "Accelerated Linearized Bregman Method" (PDF). Journal of Scientific Computing. Plenum Press (published 1 Feb 2013). 54 (2–3): 428–453. arXiv:1106.5413. doi:10.1007/s10915-012-9592-9. ISSN 0885-7474. S2CID 14781930. Retrieved 22 Apr 2021.
  10. ^ Wang, Huahua; Banerjee, Arindam (13 Jun 2013). "Bregman alternating direction method of multipliers". NIPS'14: Proceedings of the 27th International Conference on Neural Information Processing Systems. 2: 2816–2824. arXiv:1306.3203. Retrieved 20 Apr 2021.
  11. ^ Jia, Rong-Qing (3 Oct 2008). "Convergence analysis of the Bregman method for the variational model of image denoising" (PDF). Applied and Computational Harmonic Analysis (published Nov 2009). 27 (3): 367–379. doi:10.1016/j.acha.2009.05.002. Retrieved 22 Apr 2021.

bregman, method, this, article, multiple, issues, please, help, improve, discuss, these, issues, talk, page, learn, when, remove, these, template, messages, this, article, needs, additional, citations, verification, please, help, improve, this, article, adding. This article has multiple issues Please help improve it or discuss these issues on the talk page Learn how and when to remove these template messages This article needs additional citations for verification Please help improve this article by adding citations to reliable sources Unsourced material may be challenged and removed Find sources Bregman method news newspapers books scholar JSTOR April 2021 Learn how and when to remove this template message This article may be too technical for most readers to understand Please help improve it to make it understandable to non experts without removing the technical details April 2021 Learn how and when to remove this template message Learn how and when to remove this template message The Bregman method is an iterative algorithm to solve certain convex optimization problems involving regularization 1 The original version is due to Lev M Bregman who published it in 1967 2 The algorithm is a row action method accessing constraint functions one by one and the method is particularly suited for large optimization problems where constraints can be efficiently enumerated citation needed The algorithm works particularly well for regularizers such as the ℓ 1 displaystyle ell 1 norm where it converges very quickly because of an error cancellation effect 3 Contents 1 Algorithm 2 Applications 3 Generalizations and drawbacks 3 1 Linearized Bregman 3 2 Split Bregman 4 ReferencesAlgorithm EditIn order to be able to use the Bregman method one must frame the problem of interest as finding min u J u f u displaystyle min u J u f u where J displaystyle J is a regularizing function such as ℓ 1 displaystyle ell 1 3 The Bregman distance is defined as D p u v J u J v p u v displaystyle D p u v J u J v langle p u v rangle where p displaystyle p belongs to the subdifferential of J displaystyle J at u displaystyle u which we denoted J u displaystyle partial J u 3 4 One performs the iteration u k 1 min u a D u u k f u displaystyle u k 1 min u alpha D u u k f u with a displaystyle alpha a constant to be chosen by the user and the minimization performed by an ordinary convex optimization algorithm 3 or u k 1 min u D p k u u k f u displaystyle u k 1 min u D p k u u k f u with p k displaystyle p k chosen each time to be a member of J u k displaystyle partial J u k 4 The algorithm starts with a pair of primal and dual variables Then for each constraint a generalized projection onto its feasible set is performed updating both the constraint s dual variable and all primal variables for which there are non zero coefficients in the constraint functions gradient In case the objective is strictly convex and all constraint functions are convex the limit of this iterative projection converges to the optimal primal dual pair citation needed In the case of a basis pursuit type problem min x A x b x 1 1 2 a x 2 2 displaystyle min x Ax b x 1 frac 1 2 alpha x 2 2 the Bregman method is equivalent to ordinary gradient descent on the dual problem min y b t y a 2 A t y Proj 1 1 n A t y 2 displaystyle min y b t y frac alpha 2 A t y text Proj 1 1 n A t y 2 5 An exact regularization type effect also occurs in this case if a displaystyle alpha exceeds a certain threshold the optimum value of x displaystyle x is precisely the optimum solution of min x A x b x 1 displaystyle min x Ax b x 1 3 5 Applications EditThe Bregman method or its generalizations can be applied to Image deblurring or denoising 3 including total variation denoising 4 MR image clarification needed reconstruction 3 Magnetic resonance imaging 1 6 Radar 1 Hyperspectral imaging 7 Compressed sensing 5 Least absolute deviations or ℓ 1 displaystyle ell 1 regularized linear regression 8 Covariance selection learning a sparse covariance matrix 8 Matrix completion 9 Structural risk minimization 8 Generalizations and drawbacks EditThe method has links to the method of multipliers and dual ascent method through the so called Bregman alternating direction method of multipliers 10 7 generalizing the alternating direction method of multipliers 8 and multiple generalizations exist One drawback of the method is that it is only provably convergent if the objective function is strictly convex In case this can not be ensured as for linear programs or non strictly convex quadratic programs additional methods such as proximal gradient methods have been developed citation needed In the case of the Rudin Osher Fatemi model of image denoising clarification needed the Bregman method provably converges 11 Some generalizations of the Bregman method include Inverse scale space method clarification needed 3 Linearized Bregman 3 Logistic Bregman 3 Split Bregman 3 Linearized Bregman Edit In the Linearized Bregman method one linearizes the intermediate objective functions D p u u k f u displaystyle D p u u k f u by replacing the second term with f u k f u k u u k displaystyle f u k langle f u k u u k rangle which approximates the second term near u k displaystyle u k and adding the penalty term 1 2 d u u k 2 2 displaystyle frac 1 2 delta u u k 2 2 for a constant d displaystyle delta The result is much more computationally tractable especially in basis pursuit type problems 4 5 In the case of a generic basis pursuit problem min m u 1 1 2 A u f 2 2 displaystyle min mu u 1 frac 1 2 Au f 2 2 one can express the iteration as v k 1 v k A t f A u k u k 1 i d shrink v k i m displaystyle v k 1 v k A t f Au k u k 1 i delta text shrink v k i mu for each component i displaystyle i where we define shrink y a y a y a 0 y a a y a y a displaystyle text shrink y a begin cases y a amp y in a infty 0 amp y in a a y a amp y in infty a end cases 4 Sometimes when running the Linearized Bregman method there are periods of stagnation where the residual clarification needed is almost constant To alleviate this issue one can use the Linearized Bregman method with kicking where one essentially detects the beginning of a stagnation period then predicts and skips to the end of it 4 5 Since Linearized Bregman is mathematically equivalent to gradient descent it can be accelerated with methods to accelerate gradient descent such as line search L BGFS Barzilai Borwein steps or the Nesterov method the last has been proposed as the accelerated linearized Bregman method 5 9 Split Bregman Edit The Split Bregman method solves problems of the form min u F u 1 H u displaystyle min u Phi u 1 H u where F displaystyle Phi and H displaystyle H are both convex 4 particularly problems of the form min u F u 1 K u f 2 displaystyle min u Phi u 1 Ku f 2 6 We start by rewriting it as the constrained optimization problem min u d F u d 1 H u displaystyle min u d Phi u d 1 H u then relax it into min u d d 1 H u l 2 d F u 2 2 displaystyle min u d d 1 H u frac lambda 2 d Phi u 2 2 where l displaystyle lambda is a constant By defining J u d d H u displaystyle J u d d H u one reduces the problem to one that can be solved with the ordinary Bregman algorithm 4 6 The Split Bregman method has been generalized to optimization over complex numbers using Wirtinger derivatives 1 References Edit a b c d Xiong Kai Zhao Guanghui Shi Guangming Wang Yingbin 2019 09 12 A Convex Optimization Algorithm for Compressed Sensing in a Complex Domain The Complex Valued Split Bregman Method Sensors published 18 Oct 2019 19 20 4540 Bibcode 2019Senso 19 4540X doi 10 3390 s19204540 PMC 6832202 PMID 31635423 Bregman L A Relaxation Method of Finding a Common Point of Convex Sets and its Application to Problems of Optimization Dokl Akad Nauk SSSR v 171 No 5 1966 p p 1019 1022 English translation Soviet Math Dokl v 7 1966 p p 1578 1581 a b c d e f g h i j k Yin Wotao 8 Dec 2009 The Bregman Methods Review and New Results PDF Archived PDF from the original on 2010 06 13 Retrieved 16 Apr 2021 a b c d e f g h Bush Jacqueline 10 Jun 2011 University of California Santa Barbara Senior Thesis Bregman Algorithms PDF University of California Santa Barbara Archived PDF from the original on 2016 11 30 Retrieved 16 Apr 2021 a b c d e f Yin Wotao 28 May 2009 Analysis and Generalizations of the Linearized Bregman Method PDF SIAM Journal on Imaging Sciences 3 4 856 877 doi 10 1137 090760350 Retrieved 16 Apr 2021 a b c Goldstein Tom Osher Stanley 2 Jun 2008 The Split Bregman Method for L1 Regularized Problems SIAM J Imaging Sci 2 2 323 343 doi 10 1137 080725891 Retrieved 22 Apr 2021 a b Jiang Chunzhi May 2015 Comparison of Variable Penalty ADMM with Split Bregman Method on Hyperspectral Imaging Problems Archived from the original on 2020 03 23 Retrieved 20 Apr 2021 a b c d Boyd Stephen Parikh Neal Chu Eric Peleato Borja Eckstein Jonathan 19 Nov 2010 Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers Foundations and Trends in Machine Learning 3 1 122 CiteSeerX 10 1 1 722 981 doi 10 1561 2200000016 a b Huang Bo Ma Shiqian Goldfarb Donald 27 Jun 2011 Accelerated Linearized Bregman Method PDF Journal of Scientific Computing Plenum Press published 1 Feb 2013 54 2 3 428 453 arXiv 1106 5413 doi 10 1007 s10915 012 9592 9 ISSN 0885 7474 S2CID 14781930 Retrieved 22 Apr 2021 Wang Huahua Banerjee Arindam 13 Jun 2013 Bregman alternating direction method of multipliers NIPS 14 Proceedings of the 27th International Conference on Neural Information Processing Systems 2 2816 2824 arXiv 1306 3203 Retrieved 20 Apr 2021 Jia Rong Qing 3 Oct 2008 Convergence analysis of the Bregman method for the variational model of image denoising PDF Applied and Computational Harmonic Analysis published Nov 2009 27 3 367 379 doi 10 1016 j acha 2009 05 002 Retrieved 22 Apr 2021 Retrieved from https en wikipedia org w index php title Bregman method amp oldid 1121553578, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.