Skip to main content
Log in

A Globally and R-Linearly Convergent Hybrid HS and PRP Method and its Inexact Version with Applications

  • Published:
Ukrainian Mathematical Journal Aims and scope

We present a hybrid HS- and PRP-type conjugate gradient method for smooth optimization that converges globally and R-linearly for general functions. We also introduce its inexact version for problems of this kind in which gradients or values of the functions are unknown or difficult to compute. Moreover, we apply the inexact method to solve a nonsmooth convex optimization problem by converting it into a one-time continuously differentiable function by the method of Moreau–Yosida regularization.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Similar content being viewed by others

References

  1. M. Al-Baali, “Descent property and global convergence of the Fletcher–Reeves method with inexact line search,” IMA J. Numer. Anal., 5, 121–124 (1985).

    Article  MathSciNet  MATH  Google Scholar 

  2. R. H. Byrd and J. Nocedal, “A tool for the analysis of quasi-Newton methods with application to unconstrained minimization,” SIAM J. Numer. Anal., 26, 727–739 (1989).

    Article  MathSciNet  MATH  Google Scholar 

  3. A. R. Conn, K. Scheinberg, and Ph. L. Toint, “Recent progress in unconstrained nonlinear optimization without derivatives,” Math. Program., 79, 397–414 (1997).

    MathSciNet  MATH  Google Scholar 

  4. Y. Dai, “Nonlinear conjugate gradient methods,” http://lsec.cc.ac.cn/dyh/worklist.html.

  5. Y. Dai and Y. Yuan, “An efficient hybrid conjugate gradient method for unconstrained optimization,” Ann. Oper. Res., 103, 33–47 (2001).

    Article  MathSciNet  MATH  Google Scholar 

  6. Y. Dai and L. Z. Liao, “New conjugacy conditions and related nonlinear conjugate gradient methods,” Appl. Math. Optim., 43, 87–101 (2001).

    Article  MathSciNet  MATH  Google Scholar 

  7. Y. Dai and Y. Yuan, “A nonlinear conjugate gradient method with a strong global convergence property,” SIAM J. Optim., 10, 177–182 (2000).

    Article  Google Scholar 

  8. J. E. Dennis and J. J. More, “A characterization of superlinear convergence and its applications to quasi-Newton methods,” Math. Comput., 28, 549–560 (1974).

    Article  MathSciNet  MATH  Google Scholar 

  9. R. Fletcher and C. Reeves, “Function minimization by conjugate gradients,” Comput. J., 7, 149–154 (1964).

    Article  MathSciNet  MATH  Google Scholar 

  10. M. Fukushima and L. Qi, “A globally and superlinearly convergent algorithm for nonsmooth convex minimization,” SIAM J. Optim., 6, 1106–1120 (1996).

    Article  MathSciNet  MATH  Google Scholar 

  11. W.W. Hager and H. Zhang, “A new conjugate gradient method with guaranteed descent and an efficient line search,” SIAM J. Optim., 16, 170–192 (2005).

    Article  MathSciNet  MATH  Google Scholar 

  12. M. R. Hestenes and E. L. Stiefel, “Method of conjugate gradient for solving linear systems,” J. Res. Nat. Bur. Stand., 49, 409–432 (1952).

    Article  MathSciNet  MATH  Google Scholar 

  13. J. C. Gilbert and J. Nocedal, “Global convergence properties of conjugate gradient methods for optimization,” SIAM J. Optim., 2, 21–42 (1992).

    Article  MathSciNet  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Additional information

Published in Ukrains’kyi Matematychnyi Zhurnal, Vol. 67, No. 6, pp. 752–762, June, 2015.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhou, W. A Globally and R-Linearly Convergent Hybrid HS and PRP Method and its Inexact Version with Applications. Ukr Math J 67, 853–865 (2015). https://doi.org/10.1007/s11253-015-1118-9

Download citation

  • Received:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11253-015-1118-9

Keywords

Navigation