A new class of improved convex underestimators for twice continuously differentiable constrained NLPs

Ioannis G. Akrotirianakis, Christodoulos A. Floudas

Research output: Contribution to journalArticlepeer-review

57 Scopus citations

Abstract

We present a new class of convex underestimators for arbitrarily nonconvex and twice continuously differentiable functions. The underestimators are derived by augmenting the original nonconvex function by a nonlinear relaxation function. The relaxation function is a separable convex function, that involves the sum of univariate parametric exponential functions. An efficient procedure that finds the appropriate values for those parameters is developed. This procedure uses interval arithmetic extensively in order to verify whether the new underestimator is convex. For arbitrarily nonconvex functions it is shown that these convex underestimators are tighter than those generated by the αBB method. Computational studies complemented with geometrical interpretations demonstrate the potential benefits of the proposed improved convex underestimators.

Original languageEnglish (US)
Pages (from-to)367-390
Number of pages24
JournalJournal of Global Optimization
Volume30
Issue number4
DOIs
StatePublished - Dec 2004
Externally publishedYes

All Science Journal Classification (ASJC) codes

  • Business, Management and Accounting (miscellaneous)
  • Computer Science Applications
  • Control and Optimization
  • Management Science and Operations Research
  • Applied Mathematics

Keywords

  • αBB
  • Convex underestimators
  • Global optimization

Fingerprint

Dive into the research topics of 'A new class of improved convex underestimators for twice continuously differentiable constrained NLPs'. Together they form a unique fingerprint.

Cite this