/OptimPack

OptimPack is a library for large optimization problems.

Primary LanguageCOtherNOASSERTION

OptimPack

(version 3.0.0)

Travis build status

This is OptimPack, a library for solving optimization problems. The library is mostly targeted at vary large problems (e.g. as the ones encountered in image restoration) but also provide routines for problems of smaller size.

This document provides a general overview of OptimPack, for more specific information, see:

Large scale problems

For large scale problems involving millions of variables (or more), OptimPack provides:

  • several non-linear conjugate gradient (NLCG) methods (see refs. [1-3]);

  • limited memory variable metric (LBFGS, see ref. [4]) possibly with bound constraints and/or preconditioning (VMLMB, see ref. [5], or BLMVM, see ref. [6]);

  • inexact monotone and nonmonotone line searches (see ref. [7,8]);

  • linear conjugate gradients [1].

Most of the documentation is in the header files, e.g. src/optimpack.h, in Doxygen format.

The large scale optimizers of the OptimPack library can work with the unknowns stored in almost any form (providing a minimal set of functions to manipulate them are implemented). This feature may be used to exploit hardware acceleration, multi-threading or to distribute the storage and computation across multiple machines.

Problems of small to moderate size

For problems of small to moderate size, OptimPack provides:

  • Moré & Sorensen method to compute a trust region step (see ref. [13]);

  • Mike Powell's COBYLA (see ref. [10]), NEWUOA (see ref. [11]), and BOBYQA (see ref. [12]) algorithms for minimizing a function of many variables. These methods are derivatives free (only the function values are needed). NEWUOA is for unconstrained optimization. COBYLA accounts for general inequality constraints. BOBYQA accounts for bound constraints on the variables.

  • Brent's method for the minimization of an univariate function.

OptimPack Bindings

OptimPack library is written in C but, in order to make embedding OptimPack into another language as easy as possible, the routines use reverse communication: all local variables needed by the optimizers get saved into workspaces created by the library and the optimizers never explicitely call the penalty function to optimize.

The following language bindings allow OptimPack to be used in other programming languages:

References

  1. M.R. Hestenes & E. Stiefel, "Methods of Conjugate Gradients for Solving Linear Systems," Journal of Research of the National Bureau of Standards 49, 409-436 (1952).

  2. W.W. Hager & H. Zhang, "A survey of nonlinear conjugate gradient methods," Pacific Journal of Optimization, Vol. 2, pp. 35-58 (2006).

  3. W.W. Hager & H. Zhang, "A New Conjugate Gradient Method with Guaranteed Descent and an Efficient Line Search," SIAM J. Optim., Vol. 16, pp. 170-192 (2005).

  4. D. Liu and J. Nocedal, "On the limited memory BFGS method for large scale optimization", Mathematical Programming B 45, 503-528 (1989).

  5. É. Thiébaut, "Optimization issues in blind deconvolution algorithms," in Astronomical Data Analysis II, SPIE Proc. 4847, 174-183 (2002).

  6. S.J. Benson & J.J. Moré, "A limited memory variable metric method in subspaces and bound constrained optimization problems", in Subspaces and Bound Constrained Optimization Problems, (2001).

  7. E.G. Birgin, J.M. Martínez & M. Raydan, "Nonmonotone Spectral Projected Gradient Methods on Convex Sets," SIAM J. Optim. 10, 1196-1211 (2000).

  8. Jorge J. Moré and David J. Thuente, "Line search algorithms with guaranteed sufficient decrease" in ACM Transactions on Mathematical Software (TOMS) Volume 20, Issue 3, Pages 286-307 (September 1994).

  9. T. Steihaug, "The conjugate gradient method and trust regions in large scale optimization", SIAM Journal on Numerical Analysis, vol. 20, pp. 626-637, 1983.

  10. M.J.D. Powell, "A direct search optimization method that models the objective and constraint functions by linear interpolation," in Advances in Optimization and Numerical Analysis Mathematics and Its Applications, vol. 275 (eds. Susana Gomez and Jean-Pierre Hennart), Kluwer Academic Publishers, pp. 51-67 (1994).

  11. M.J.D. Powell, "The NEWUOA software for unconstrained minimization without derivatives", in Large-Scale Nonlinear Optimization, editors G. Di Pillo and M. Roma, Springer (2006), pages 255-297.

  12. M.J.D. Powell, "The BOBYQA Algorithm for Bound Constrained Optimization Without Derivatives." Technical report, Department of Applied Mathematics and Theoretical Physics, University of Cambridge (2009).

  13. J.J. Moré & D.C. Sorensen, "Computing A Trust Region Step," SIAM J. Sci. Stat. Comp. 4, 553-572 (1983).

  14. R.P. Brent, "Algorithms for Minimization without Derivatives," Prentice-Hall, Inc. (1973).

Authors

Credits

The development of OptimPack was supported by the MiTiV project funded by the French Agence Nationale pour la Recherche (ref. ANR-09-EMER-008).

The older version of OptimPack is available here***

License

The OptimPack library is released under the MIT "Expat" License.