Global optimization

Source: Wikipedia, the free encyclopedia.
(Redirected from
Global optimisation
)

Global optimization is a branch of

minima or maxima
of a function or a set of functions on a given set. It is usually described as a minimization problem because the maximization of the real-valued function is equivalent to the minimization of the function .

Given a possibly nonlinear and non-convex continuous function with the global minima and the set of all global minimizers in , the standard minimization problem can be given as

that is, finding and a global minimizer in ; where is a (not necessarily convex) compact set defined by inequalities .

Global optimization is distinguished from local optimization by its focus on finding the minimum or maximum over the given set, as opposed to finding local minima or maxima. Finding an arbitrary local minimum is relatively straightforward by using classical local optimization methods. Finding the global minimum of a function is far more difficult: analytical methods are frequently not applicable, and the use of numerical solution strategies often leads to very hard challenges.

Applications

Typical examples of global optimization applications include:

  • Protein structure prediction (minimize the energy/free energy function)
  • Computational phylogenetics (e.g., minimize the number of character transformations in the tree)
  • Traveling salesman problem
    and electrical circuit design (minimize the path length)
  • Chemical engineering (e.g., analyzing the Gibbs energy)
  • Safety verification, safety engineering (e.g., of mechanical structures, buildings)
  • Worst-case analysis
  • Mathematical problems (e.g., the Kepler conjecture)
  • Object packing (configuration design) problems
  • The starting point of several molecular dynamics simulations consists of an initial optimization of the energy of the system to be simulated.
  • Spin glasses
  • Calibration of
    radio propagation models
    and of many other models in the sciences and engineering
  • Curve fitting like non-linear least squares analysis and other generalizations, used in fitting model parameters to experimental data in chemistry, physics, biology, economics, finance, medicine, astronomy, engineering.
  • IMRT radiation therapy planning

Deterministic methods

The most successful general exact strategies are:

Inner and outer approximation

In both of these strategies, the set over which a function is to be optimized is approximated by polyhedra. In inner approximation, the polyhedra are contained in the set, while in outer approximation, the polyhedra contain the set.

Cutting-plane methods

The cutting-plane method is an umbrella term for optimization methods which iteratively refine a

mixed integer linear programming (MILP) problems, as well as to solve general, not necessarily differentiable convex optimization problems. The use of cutting planes to solve MILP was introduced by Ralph E. Gomory and Václav Chvátal
.

Branch and bound methods

Branch and bound (BB or B&B) is an algorithm design paradigm for discrete and combinatorial optimization problems. A branch-and-bound algorithm consists of a systematic enumeration of candidate solutions by means of state space search: the set of candidate solutions is thought of as forming a rooted tree with the full set at the root. The algorithm explores branches of this tree, which represent subsets of the solution set. Before enumerating the candidate solutions of a branch, the branch is checked against upper and lower estimated bounds on the optimal solution, and is discarded if it cannot produce a better solution than the best one found so far by the algorithm.

Interval methods

Interval arithmetic, interval mathematics, interval analysis, or interval computation, is a method developed by mathematicians since the 1950s and 1960s as an approach to putting bounds on

numerical methods
that yield reliable results. Interval arithmetic helps find reliable and guaranteed solutions to equations and optimization problems.

Methods based on real algebraic geometry

Real algebra is the part of algebra which is relevant to real algebraic (and semialgebraic) geometry. It is mostly concerned with the study of ordered fields and ordered rings (in particular real closed fields) and their applications to the study of positive polynomials and sums-of-squares of polynomials. It can be used in convex optimization

Stochastic methods

Several exact or inexact Monte-Carlo-based algorithms exist:

Direct Monte-Carlo sampling

In this method, random simulations are used to find an approximate solution.

Example: The

traveling salesman problem
is what is called a conventional optimization problem. That is, all the facts (distances between each destination point) needed to determine the optimal path to follow are known with certainty and the goal is to run through the possible travel choices to come up with the one with the lowest total distance. However, let's assume that instead of wanting to minimize the total distance traveled to visit each desired destination, we wanted to minimize the total time needed to reach each destination. This goes beyond conventional optimization since travel time is inherently uncertain (traffic jams, time of day, etc.). As a result, to determine our optimal path we would want to use simulation - optimization to first understand the range of potential times it could take to go from one point to another (represented by a probability distribution in this case rather than a specific distance) and then optimize our travel decisions to identify the best path to follow taking that uncertainty into account.

Stochastic tunneling

Stochastic tunneling (STUN) is an approach to global optimization based on the Monte Carlo method-sampling of the function to be objectively minimized in which the function is nonlinearly transformed to allow for easier tunneling among regions containing function minima. Easier tunneling allows for faster exploration of sample space and faster convergence to a good solution.

Parallel tempering

Parallel tempering, also known as replica exchange MCMC sampling, is a simulation method aimed at improving the dynamic properties of Monte Carlo method simulations of physical systems, and of Markov chain Monte Carlo (MCMC) sampling methods more generally. The replica exchange method was originally devised by Swendsen,[1] then extended by Geyer[2] and later developed, among others, by Giorgio Parisi.,[3][4] Sugita and Okamoto formulated a molecular dynamics version of parallel tempering:[5] this is usually known as replica-exchange molecular dynamics or REMD.

Essentially, one runs N copies of the system, randomly initialized, at different temperatures. Then, based on the Metropolis criterion one exchanges configurations at different temperatures. The idea of this method is to make configurations at high temperatures available to the simulations at low temperatures and vice versa. This results in a very robust ensemble which is able to sample both low and high energy configurations. In this way, thermodynamical properties such as the specific heat, which is in general not well computed in the canonical ensemble, can be computed with great precision.

Heuristics and metaheuristics

Other approaches include heuristic strategies to search the search space in a more or less intelligent way, including:

  • Ant colony optimization (ACO)
  • Simulated annealing, a generic probabilistic metaheuristic
  • Tabu search, an extension of local search capable of escaping from local minima
  • evolution strategies
    )
  • candidate solution
    with regard to a given measure of quality
  • ant colony optimization
    )
  • Memetic algorithms, combining global and local search strategies
  • Reactive search optimization (i.e. integration of sub-symbolic machine learning techniques into search heuristics)
  • Graduated optimization, a technique that attempts to solve a difficult optimization problem by initially solving a greatly simplified problem, and progressively transforming that problem (while optimizing) until it is equivalent to the difficult optimization problem.[6][7][8]

Response surface methodology-based approaches

See also

Footnotes

  1. ^ Swendsen RH and Wang JS (1986) Replica Monte Carlo simulation of spin glasses Physical Review Letters 57 : 2607–2609
  2. ^ C. J. Geyer, (1991) in Computing Science and Statistics, Proceedings of the 23rd Symposium on the Interface, American Statistical Association, New York, p. 156.
  3. S2CID 13963102
    .
  4. ^ David J. Earl and Michael W. Deem (2005) "Parallel tempering: Theory, applications, and new perspectives", Phys. Chem. Chem. Phys., 7, 3910
  5. .
  6. ^ Thacker, Neil; Cootes, Tim (1996). "Graduated Non-Convexity and Multi-Resolution Optimization Methods". Vision Through Optimization.
  7. ]
  8. ^ Hossein Mobahi, John W. Fisher III. On the Link Between Gaussian Homotopy Continuation and Convex Envelopes, In Lecture Notes in Computer Science (EMMCVPR 2015), Springer, 2015.
  9. ^ Jonas Mockus (2013). Bayesian approach to global optimization: theory and applications. Kluwer Academic.

References

Deterministic global optimization:

For simulated annealing:

  • Kirkpatrick, S.; Gelatt, C. D.; Vecchi, M. P. (1983-05-13). "Optimization by Simulated Annealing". Science. 220 (4598). American Association for the Advancement of Science (AAAS): 671–680.
    S2CID 205939
    .

For reactive search optimization:

For stochastic methods:

For parallel tempering:

For continuation methods:

For general considerations on the dimensionality of the domain of definition of the objective function:

  • Hamacher, Kay (2005). "On stochastic global optimization of one-dimensional functions". Physica A: Statistical Mechanics and Its Applications. 354. Elsevier BV: 547–557.
    ISSN 0378-4371
    .

For strategies allowing one to compare deterministic and stochastic global optimization methods

External links