 American Journal of Computational Mathematics, 2011, 1, 240-246 doi:10.4236/ajcm.2011.14028 Published Online December 2011 (http://www.SciRP.org/journal/ajcm) Copyright © 2011 SciRes. AJCM A Look at the Tool of BYRD and NOCEDAL* Linghua Huang1, Guoyin Li2, Gonglin Yuan3 1Department of Mathematics and Statistics, Guangxi University of Finance and Economics, Nanning, China 2Department of Applied Mathematics, The University of New South Wales, Sydney, Australia 3Department of Mathematics and Information Science, Guangxi University, Nanning, China E-mail: linghuahuang@163. com, g.li@unsw.edu.au, glyuan@gxu.edu.cn Received July 13, 2011; revised September 6, 2011; accepted September 15, 2011 Abstract A power tool for the analysis of quasi-Newton methods has been proposed by Byrd and Nocedal ([1], 1989). The purpose of this paper is to make a study to the basic property (BP) given in [1]. As a result of the BP, a sufficient condition of global convergence for a class of quasi-Newton methods for solving unconstrained minimization problems without convexity assumption is given. A modified BFGS formula is designed to match the requirements of the sufficient condition. The numerical results show that the proposed method is very encouraging. Keywords: quasi-Newton Method, Unconstrained Minimization, Nonconvex Problem, Global Convergence 1. Introduction Given a real-valued function :n RR, we are in- terested in solving the unconstrained optimization pro- blem min n xx R by Quasi-Newton methods, which have the form 1, kkkk xd where k is a steplength, and is a search direction given by the following linear equation k d 0. kk Bd g (1.1) The matrix k is updated at every step such that satisfies the so-called secant equation Bk B 1, kk k Bs where k1kk xx , 1kk k g and denotes the gradient of at . Global convergence of quasi-Newton methods has been widely studied in the past two decades. For convex minimization, Powell [2] showed that, with the weak Wolfe-Powell line search strategies, lim inf0. kk g Werner [3] made an extension of Powell’s result to some other line searches. Byrd, Nocedal and Yuan [4] made an inside study for the restricted Broyden class of quasi- Newton methods. Byrd and Nocedal (1989) proposed a very useful tool for the analysis of quasi-Newton meth- ods. The basic property (BP) given by Byrd and Nocedal (1989) characterized not only the BFGS formula but also any formula with the structure of the BFGS, some of the examples are the modified BFGS methods given by Li and Fukushima [5-7]. In [5], Li and Fukushima gave a modified BFGS method with good convergence proper- ties for solving symmetric nonlinear equations, while in [6,7], the BFGS type methods with global and superlin- ear convergence are designed for nonconvex optimiza- tion problems. The proofs of some main results given in [5-7] are related closely to the BP. Some modified BFGS methods which possess not only the gradient value but also the function value information have been proposed (see [8,9] etc.). The main purpose of this paper is to give some insight of the BP for a class of quasi-Newton me- thods. In the next section we recall some basic concepts and results of [1], and then study a class of quasi-Newton methods. By using the BP, we obtain a natural property of the proposed methods. Moreover, we give a sufficient condition for the quasi-Newton methods, which is moti- vated also by BP. In section 3, we design a modified BFGS method to match the requirements of the given sufficient condition. As we will see, the proposed method *This work is supported by Program for Excellent Talents in Guangxi Higher Education Institutions, China NSF grands 10761001, Guangxi Education research project grands 201012MS013, Guangxi SF grands 0991028, and the Scientific Research Foundation of GuangXi Univer- sity (Grant No. XJZ110632).
 L. H. HUANG ET AL.241 is globally and superlinearly convergent for nonconvex unconstrained minimization problems. The numerical results are contained in Section 4. In Section 5, we study the set of good iterates of BFGS and DFP formulas with different step size strategies by empirical analysis. Throughout this paper, the norm is the Euclidean vector norm. 2. Set of Good Iterates After making an inside study on the tool given by Byrd and Nocedal, we found that the main contribution of [1] are three: 1) gave a power tool for analysis of quasi- Newton methods; 2) showed the BFGS formula pos- sesses the BP that is independent of the algorithmic con- text of the update for convex minimization proplems and 3) characterized the set of good iterates by using the in- formation of the update matrix and the iteration point k B k . For the following BFGS formula 1, TT kkk kkk kk TT kkk kk Bss Bvv BB Bsv s (2.1) Byrd and Nocedal [1] proposed a basic property that indicates, under some conditions, the most iterates gen- erated by (2.1) are good iterates. It is more interesting to note that the above conclusion is independent on any line search strategies. The BP given by Byrd and Nocedal (Theorem 2.1 of [1] ) is as follows: Theorem 2.1. Let be generated by (2.1) with the following properties: and for all , k B 1 B01k 1 T kk T kk vs ss for some 10, (2.2) 2 2 k T kk v vs for some 20, (2.3) Then for any there exist constants 0,1p 123 ,, 0 such that, for all , the relations 1k 1 jj , s B T jjj j sBs d (2.4) 2, jjj T jj sBs ss 3 (2.5) 3 2 1 , jj j Bs s (2.6) hold for at least values of k p 1, 2, 3,,.jk The conclusion of Theorem 2.1 is right for any for- mula with the form 1 TT kkkkk k kk T kkk kk BrrBuu BB rBr ur T (2.7) if 1 and for all, kk , where kk 0Bk0 T ur,n ur . Moreover, the proof of the conclusion for (2.7) does not need to change anywhere. This makes one can choose k and k such that k, (2.2) and (2.3) hold. In fact, Li and Fukushima have followed this way and gave some modified BFGS formula which possess global and superlinear convergence for nonconvex minimization [6] and for symmetric nonlinear equations [5]. In this sense, the tool given in [1] for proving Theorem 2.1 is very powerful. y0B Let 0, , all nonngative integerN. Define four functions , , and SGI 1 2 as follows: 123 ::2hastheform, alltheindex which satisfis2.4,2.5 and2.6simultaneeousl N SGISGI , 11k ::2 has the form ()B; N kk ckdcd 22 2 12 1122 ::2 has the form () ; ::2 has the form , . N T kkkk N c kcddBd cc cc In [1], the set ,,SGI 123 was been called the “set of good iterates”. From Theorem 2.1 and p can be chosen to be close to 1, we can deduce that, for any line search strategies, if the conditions (2.2) and (2.3) are satisfied, then most of the iterates given by the BFGS formula are good iterates. The meaning of “good” is cer- tified in [1] (Theorem 3.1) by proving, with some certain line search strategies, that any quasi-Newton method is R-linearly convergent for uniform convex minimization problems. In the remainder of this section, we will give a “set of good iterates” for a general quasi-Newton methods. In order to simplify the presentation, we use 00Bto denote any nn symmetric and positive semi-definite (definite) matrix B. We state the method, which will be called the GQNLSS: (general quasi-Newton method with some certain line search strategies) as follows. Algorithm 2.1: GQNLSS Step 0: Choose an initial point and an initial 1 n x matrix . Choose 10B 1 0, ,0,1 2 Set :1k . Copyright © 2011 SciRes. AJCM
 L. H. HUANG ET AL. 242 Step 1: If , Stop. 0 k g Step 2: Solve (1.1) to obtain a search direction . k d Step 3: find k by some certain line search strate- gies. Step 4: Set 1kkkk xd . Update by some formula. 10 k B Step 5: Set and go to Step 1. :kk1 The line search strategies used in the GQNLSS is one of the following three forms: 1) Efficient line search: find k satisfies 2 12, T kk kkk k k gd fxd fx d where η1 is a positive constant. 2) Standard Armijo line search: find k k j such that is the smallest nonnegative integer satisfy- ing k j 2, jkjk T kk kkk xdfx gd (2.8) where and are constants. 0,1 2 3) Weak Wolfe-Powell (WWP) line searches: find 0,1 k satisfies condition: 2, jkjk T kk kkk xdfx gd (2.9) and T kkk kk xd g d (2.10) where and are constants. 0,1 3 3 In order to study the convergence behavior of Algo- rithm 2.1, we will impose the following two assumptions, which have been widely used in the literature to analyze the global convergence of iterative solution methods for minimization problem with inexact line searches (see [10,11] etc.). ,1 Assumption A. The level set is bounded. 1 n Rfxfx Assumption B. There exists a constant L such that for any , ,xy xgy Lxy. The following natural property, characterizes the up- dated matrix k and the direction generated by GQNLLS method. Bk d Theorem 2.2. Suppose that Assumptions A and B hold, ,,, kkkk Bd is generated by GQNLSS. Then 2 2 1 . T kkk kk dBd d (2.11) Proof: From Assumption A, we have that . Thus k x 1 1 kk k fx fx (2.12a) 1) For the line searches 1), we have (2.11) by using (1.1), and (2.12). 0B k 2) For the line searches 2), it suffices to consider the case 1 k . From the definition of k , we have 2 T kk kkk kk xdfx g d . Using the Mean Value Theorem in the above inequal- ity, we obtain 0,1 k , such that 2. TT kkk kkkk kk xdd gd Dividing the both side of the above inequality by k , we have 2. TT kkkk kkk xdd gd Subtracting k T k d on the both sides of the above ine- quality, we obtain 2 1, TT kkkkk kkk xdgxd gd which combining with Assumption B yields 2 2 1. T kk kkk Ld g d Therefore 2 2 1T kk k kk xd Ld Hence, we have, by using , that 0,1 k 2 2 1T kk k k xd Ld Thus 2 2 1 T kkk kk dBd d (2.12b) by using (2.8), (1.1) and (2.12). 3) From Assumption B and (2.10), we obtain 2 1 1, T T kkkkkk k gdgg d Ld which implies that 2 1T kk k k d Ld Thus 2 1T kkk k k dBd Ld by using (1.1). Thus, (2.11) holds by using (2.12).From Copyright © 2011 SciRes. AJCM
 L. H. HUANG ET AL.243 the results of a)-c), we have (2.11). The proof is com- plete. Let k12 0kk n () k cond B be the eigenvalues of k and be the condition number of , i.e., B k B 1 nk k k cond B Theorem 2.3. Suppose that Assumptions A and B hold, ,,, kkkk Bd k is generated by GQNLSS. If there exist a positive constant M and an infinite index set such that for all , K K , k cond BM (2.13) then lim 0. k kgx (2.14) Proof: From Theorem 2.2, we have 2 2 12 limlim 0. T kkk kk kk k dBd d d Thus, 1 lim 0. kk kd Therefore, from (2.13), we obtain 1 limlimlim( )0. kknk kkkk kK kKkK B ddcondBd which implies that (2.14) holds by using (1.1). The proof is complete. From Theorem 2.3, we have the following result, which indicates that if GQNLSS fails to converge, then the condition numbers sequence will tend to infinite. k cond B Corollary 2.1. Suppose that Assumptions A and B hold, ,,, kkkk Bd is generated by GQNLSS. If liminf 0 k kgx Then lim k kcond B We call GQNLSS has property SC∞ if 12 12 , for some c,ccc (2.15) Theorem 2.4 Suppose that Assumptions A and B hold, ,,, kkkk Bd is generated by GQNLSS. If GQNLSS has property SC∞, then 12 (, ) lim 0 k kcc gx Proof: From the definition of Φ, we have 2 12 12 and c ,for ,. kk kk T kkk Bdc dd dBd k cc (2.16) From Theorem 2.1, we have 2222 22 24 22 22 2 22 0 limlimlim T kkk k k kckc kc kk dBdcd cd dd By using the definitions of and ,we obtain 1 2 12 , lim 0. kk kcc Bd Therefore, (2.16) follows (1.1). The proof is complete. The main contribution of Theorem 2.4 is that it identi- fies the convergence indices of k . It indicates that 12 , is the “set of good iterates” in the sense that the method is globally convergent. From Theorem 2.4, we see that (2.15) is a sufficient condition of the global convergence for GQNLSS. The sufficient condition is tight under the sense that the set 12 , is as same as the convergence indices of the sequence k . From the fact that ,SCI 12 123 ,, , we see that, for GQNLSS, the “set of good iterates” is little larger than which given in [1]. Note that the approach here without any convexity assumption. For BFGS formula, whether (2.2) and (2.3) hold is still open for nonconvex minimization problems. In general, it is very hard to prove them. By Theorem 2.3 and 2.4, we can obtain that if one can prove that for some positive constants and , 1 c2 c 12 1 ,k cck condBc , then the BFGS formula with any one of the three line search strategies mentioned above is globally convergent. This result can be extended to any quasi-Newton formula (see GQNLSS), such as DFP formula. 3. Design Methods with SC∞ The purpose of this section is to discuss how to design a BFGS-type method (i.e., the update has the form (2.7)) with global convergence (if possible with superlinear convergence). From Theorem 2.4, it suffices to design the methods which possess the property SC∞. Clearly, it is more than enough to find k such that for all , (2.2) and (2.3) hold by using Theorem 2.1. Although following this way may yield some strict conditions (it seems that, for nonconvex minimization problems, it excludes the BFGS update), it is better than nothing at this moment. First we will propose a BFGS-type update to match the requirements of (2.2) and (2.3), and prove the corresponding method possesses the property SC∞ by using the same way given in the proof of Theorem 2.1 in [1]. Some other new formulas will also be given then in this section. y k The first modified BFGS update is as follows: 1, TT kkkkk k kk TT kkk kk Bss Byy BB Bsys (3.1) Copyright © 2011 SciRes. AJCM
 L. H. HUANG ET AL. 244 k where 1kkkk xx d kk ms m , , . 1kk vg g k kk yv The parameter is defined by k 12 T kk kT kk vs m s where and 10, 21, are two constants. For (3.1), we have a corresponding method, which is described as follows. Algorithm 3.1: A BFGS-Type Method with WWP (BFGSTWWP) Step 0: Choose an initial point and an initial matrix . Set k := 1. 1 n x 1 Step 1: If , Stop. 0B g0 k Step 2: Solve (1.1) to obtain a search direction dk. Step 3: find k by WWP (2.9) and (2.10). Moreover, if 1 k satisfies WWP, set 1 k . Step 4: Set 1kkkk xd . Update by (3.1). 1k B Step 5: Set k := k + 1 and go to Step 1. The following Lemma shows that the updated matrix , so we can always solve (1.1) uniquely. 0 k B Lemma 3.1. Suppose that ,,, kkkk Bd 1, k kB k0B is gener- ated by BFGSTWWP. Then for all . 0 Proof: Suppose that for a given , . From the definitions of and k k yk , we have 1 1 10 TTT T kkkkkk kkk k T kk kk vsg dgdgd dBd T By using (2.10) and (1.1). Thus 12 12 1 11 TT TT kk kkkkkk T kkkk kk T kk ys vsssvs sd ss Bd (3.2) It is easy to prove that 1 by using and (3.2). By using and the induction, we may deduce that for all , . 0 k B 0 0 k B 10B Bkk Theorem 3.1. Suppose that Assumptions A and B hold, ,,, kkkk Bd is generated by BFGSTWWP. Then for any there exist constants 0,1p 12 ,0 such that, for all , the relations 1k 1 j Bs s j (3.3) 2, TT jjjj ssBs (3.4) hold for at least values of k Proof: For any given , from the definition of v and Assumption B, we have p 1, 2,,.jk kk 1, 0 kkk k T kk vggLs vs and 12 12 . kk kT kk vs mL ss Using the above two inequalities and the definitions of , and k yk vk , we have 2 22 2 2 2 12 12 2 2. TT T kkkkkkk kkkkkk T kk yyvvm ss vmvsms LLLL ss Therefore, we obtain, by using (3.2), that 1 T kk T kk ys ss (3.5) and 2 2 12 12 1 2. T kk T kk LLLL yy ys (3.6) Thus, the proof follows from that of Theorem 2.1 in [1]. From Theorem 2.4 and Theorem 3.1, we have the fol- lowing global convergence result for BFGSTWWP. Theorem 3.2 Suppose that Assumptions A and B hold, ,,, kkkk Bd is generated by BFGSTWWP. Then 12 , lim 0. k kgx Proof: Omitted. From the proof of Lemma 3.1, we observe that (3.2) is not independent of the line searches, thus, (3.5) and (3.6) is also not independent of the line searches. Therefore, like the BFGS formula, when the standard Armijo line search is used to the BFGS-type formula (3.1), k cannot be guaranteed for some cases because whether the relation kk holds is still open for nonconvex pro- blems. It is possible to give a formula such that k, (3.5) and (3.6) hold without any line search strategies. For example, the in (3.1) is replaced by 0B 0B 0 T ys k m 0 1 T kk kT kk vs m s with ∈ [1, +∞). In this case, the results of Lemma 3.1, Theorem 3.1, (therefore SC∞) hold for any k 0 . This implies clearly, by using Theorem 2.4, that the result of Theorem 3.2 holds if any one of the line search strategies given in Section 2 is used. From the proof of Theorem 2.1 in [1], it is possible to give the exact values of 1 and 2 . Let 11 1 2 2 12 12 1 ln det, 2 BtrB B LLLL M Copyright © 2011 SciRes. AJCM
 L. H. HUANG ET AL.245 and 11 11ln. 1BM p 2 Let 1 denote the two solutions of the following equation 1lntt . 2 Then 1 01 . Using the above notation, we obtain 2 1 2 e and 2 2 2.e In [6], some modified BFGS methods with global con- vergence and superlinear convergence for non-convex minimization problems have been given. It is easy to check that the methods satisfy (3.4) and (3.3). Thus, the global convergence of the methods given in [6] can be easy to obtained from Theorem 3.2. Under some condi- tions, we can prove the superlinear convergence of BFGSTWWP by using the similar way of in [6]. We do not repeat the proof here. We concluded this section by proposing another for- mula which can be view as a cautious BFGS update. Let be a very small constant, define 10,1 11 2 if 0otherwise T TT kk kk kk T kkk ss ss s and 12 T kk kk T kk vs m s Then, 1 0, k m ks It can be proved for the corre- sponding BFGS-type formula with any one of the line search strategies, that all the results in this section hold because for any 1 hold even without any line search. Notice that if for some 1, then the corresponding formula deduce to the ordinary BFGS update. , TT kk kk ss . ,TT kk kk ks ss 4. Numerical Experiments In this section, we report the numerical results for the following three methods: Algorithm 3.1: The Algorithm 3.1 with 3 110 and . Where 10 210 30.1 ,0.9. Algorithm 3.2: (3.1) and (3.7) with the Armijor line search, where 3 110 v, , 1Q0.5 and 20.1. BFGS: The BFGS formula with the WWP. ere Wh30.1 , 0.9. For each test problem, the termination is 6 10 . k gx For each problem, we choose the initial matrix B1 = I, i.e., the unit matrix. Due to the roundoff error, sometimes the directions generated by the algorithms may be not descent. We then used the steepest descent direction to take place of the related direction if The detail numerical results are listed at: http://210.36.16.53:8018/publication.asp?id=46065. 14 10 . T kk gd In order to rank the iterative numerical methods, one can compute the total number of function and gradient evaluations by the formula total ,NNFmNG (4.1) where m is some integer. According to the results on automatic differentiation [12,13], the value of m can be set to 5m . That is to say, one gradient evaluation is equivalent to m number of function evaluations if auto- matic differentiation is used. As we all known the BFGS method is considered to be the most efficient quasi- Newton method. Therefore, in this part, we compare the Algorithm 3.1 and Algorithm 3.2 with the BFGS method as follows: for each testing example i, compute the total numbers of function evaluations and gradient evaluations required by the evaluated method and the S method by the formula (4.1), and denote them by jEM j total,ijNEM and ; then calculate the ratio total,i NB FGS total, total, i i i NEMj rEM jNBFGS If 0 EM j does not work for example 0, we re- place the i 0i NEMj 0 total, by a positive constant which define as follows total, 1 max:, i NEMjijS where S1 = { ,ij: method does not work for example }. j i The geometric mean of these ratios for method EMj over all the test problems is defined by 1 i iS rEMjr EMj , where S denotes the set of the test problems and |S| the number of elements in S. One advantage of the above rule is that, the comparison is relative and hence does not be dominated by a few problems for which the method requires a great deal of function evaluations and gradient Copyright © 2011 SciRes. AJCM
 L. H. HUANG ET AL. Copyright © 2011 SciRes. AJCM 246 Table 1. Performance of these Algorithms. Algorithm 3.1 Algorithm 3.2 BFGS 0.9534 1.5763 1 functions. From Table 1, we observe that the Algorithm 3.1 out- performs the BFGS method. Therefore, the Algorithm 3.1 is the most efficient algorithm among quasi-Newton algorithms for solving minimization problems for the chosen tested problems. 5. References [1] R. Byrd and J. Nocedal, “A Tool for the Analysis of Quasi-Newton Methods with Application to Unconstrained Minimization,” SIAM Journal on Numerical Analysis, Vol. 26, No. 3, 1989, pp. 727-739. doi:10.1137/0726042 [2] M. J. D. Powell, “Some Global Convergence Properties of a variable Metric Algorithm for Minimization without Exact Line Searches,” In: R.W. Cottle and C. E. Lemke, Eds., Nonlinear Programming, SIAM-AMS Proceedings, Vol. 4, American Mathematical Society, Providence, 1976, pp.53-72. [3] J. Werner, “Über die Globale Knovergenz von Variable- Metric-Verfahre mit Nichtexakter Schrittweitenbestim- mung,” Numerische Mathematik, Vol. 31, No. 3, 1978, pp. 321-334. doi:10.1007/BF01397884 [4] R. Byrd, J. Nocedal and Y. Yuan, “Global Convergence of a Class of Quasi-Newton Methods on Convex Prob- lems,” SIAM Journal on Numerical Analysis, Vol. 24, No. 5, 1987, pp. 1171-1189. doi:10.1137/0724077 [5] D. Li and M. Fukushima, “A Global and Superlinear Con- vergent Gauss-Newton-Based BFGS Method for Sym- metric Nonlinear Equations,” SIAM Journal on Numeri- cal Analysis, Vol. 37, No. 1, 1999, pp. 152-172. doi:10.1137/S0036142998335704 [6] D. Li and M. Fukushima, “A Modified BFGS Method and Its Global Convergence in Nonconvex Minimiza- tion,” Journal of Computational and Applied Mathemat- ics, Vol. 129, No. 1-2, 2001, pp. 15-35. doi:10.1016/S0377-0427(00)00540-9 [7] D. Li and M. Fukushima, “On the Global Convergence of the BFGS Method for Nonconvex Unconstrained Optimi- zation Problems,” SIAM Journal on Optimization, Vol. 11, No. 4, 2001, pp. 1054-1064. doi:10.1137/S1052623499354242 [8] Z. Wei, G. Yu, G. Yuan and Z. Lian, “The Superlinear Convergence of a Modified BFGS-Type Method for Un- constrained Optimization,” Computational Optimization and Applications, Vol. 29, No. 3, 2004, pp. 315-332. doi:10.1023/B:COAP.0000044184.25410.39 [9] G. Yuan and Z. Wei, “Convergence Analysis of a Modi- fied BFGS Method on Convex Minimizations,” Compu- tational Optimization and Applications, Vol. 47, No. 2, 2010, pp. 237-255. doi:10.1007/s10589-008-9219-0 [10] E. G. Birgin and J. M. Martínez, “Structured Minimal- Memory Inexact Quasi-Newton Method and Secant Pre- conditioners for Augmented Lagrangian Optimization,” Computational Optimization and Applications, Vol. 39, No. 1, 2008, pp. 1-16. doi:10.1007/s10589-007-9050-z [11] G. Yuan and Z. Wei, “The Superlinear Convergence Ana- lysis of a Nonmonotone BFGS Algorithm on Convex Ob- jective Functions,” Acta Mathematics Sinica, Vol. 24, No. 1, 2008, pp. 35-42. doi:10.1007/s10114-007-1012-y [12] A. Griewank, “On Automatic Differentiation,” In: M. Iri and K. Tanabe, Eds., Mathematical Programming: Re- cent Developments and Applications, Academic Publish- ers, Cambridge, 1989, pp. 84-108. [13] Y. Dai and Q. Ni, “Testing Different Conjugate Gradient Methods for Large-Scale Unconstrained Optimization,” Journal of Computational Mathematics, Vol. 21, No. 3, 2003, pp. 311-320.
|