In the 2003 conference proceedings Celebrating the 50th Anniversary of the Metropolis Algorithm^{[9]}^{[10]}, Marshall Rosenbluth describes the algorithm in the following beautifully concise and clear manner ^{[11]}:

A simple way to do this [sampling configurations with the Boltzmann weight], as emerged after discussions with Teller, would be to make a trial move: if it decreased the energy of the system, allow it; if it increased the energy, allow it with probability exp(−ΔE/kT) as determined by a comparison with a random number. Each step, after an initial annealing period, is counted as a member of the ensemble, and the appropriate ensemble average of any quantity determined.

The control flow of the algorithm is determined by two nested loops, the outer loop over decreasing temperature simulates the cooling, and an inner loop times n Monte Carlo iterations. Each time a randomly picked neighbor state inside the inner loop provides a better energy or fitness than the current state, the neighbor becomes the new current and even new optimum if fitter than fittest so far. Otherwise, if the neighbor fitness does not exceed current, it might still become current depending on the positive fitness or energy difference ΔE, and absolute temperature T, with a probability p according to the Boltzmann factor:

where k the Boltzmann constant, and e base of the exponential function whose negative exponent ensures the [0, 1] probability interval. Accepting worse solutions is a primary feature of SA, and important to stop greedyexploitation a local optimum but to explore other areas - higher temperatures favor exploration, while decreasing temperatures make the algorithm to behave greedier in favoring exploitation of the hopefully global optimum.

Animation

Simulated annealing - searching for a maximum. ^{[16]}
With the high temprature, the numerous local maxima are left quickly through the strong noise movement -
but the global maximum is reliably found because of cooling temperature is no longer sufficient to leave it.

Pseudo Code

The C like pseudo code is based on Peter Mysliwietz' description as given in his Ph.D. thesis ^{[17]}. Several neighbor functions used to modify the weight vector were tried, where one randomly chosen element changed randomly performed well. The fitness function inside the inner loop is of course the most time consuming part. For Zugzwang, Mysliwietz used a database of 500 test-positions with a search depth of one ply, which took about three minutes on a T 800 Transputer per iteration - the higher the hit rate of found expert moves, the fitter. The whole optimization used a tHight to tLow ratio of 100, a reduction factor r of 0.95, and n=40 inner iterations.

/**
* simulatedAnnealing
* @author Peter Mysliwietz, slightly modified
* @param tHigh is the start temperature
* @param tLow is the minimal end temperature
* @param r is the temperature reduction factor < 1.0
* @param n number of iterations for each temperature
* @return best weight vector
*/
vector simulatedAnnealing(double tHigh, double tLow, double r, int n){
vector currentWeights = randomWeights();
vector bestWeights = currentWeights;double fittest = fitness(currentWeights);for(double t = tHigh; t > tLow; t *= r){for(int i =0; i < n;++i){
vector neighborWeights = neighbor(currentWeights);if( fitness(neighborWeights )> fitness(currentWeights)){
currentWeights = neighborWeights;if( fitness(neighborWeights )> fittest ){
fittest = fitness(neighborWeights);
bestWeights = neighborWeights;}}elseif(accept( fitness(currentWeights)- fitness(neighborWeights ), t)){
currentWeights = neighborWeights;}}/* for i */}/* for t */return bestWeights;}/**
* accept
* @param d is the energy difference >= 0
* @param t is the current temperature
* @return true with probability of Boltzmann factor e^(-d/kt)
*/bool accept(double d, double t ){constdouble k = 1.38064852e−23;/* joule / kelvin */double p =exp(-d /(k*t));double r =rand()/(RAND_MAX+1.0);return r < p;}

^Peter Mysliwietz (1994). Konstruktion und Optimierung von Bewertungsfunktionen beim Schach. Ph.D. thesis, 7.4. Simulated Annealing, 7.4.2. Beschreibung des Algorithmus, Abb. 29, pp. 146 (German)

Home * Programming * Algorithms * Simulated AnnealingSimulated Annealing, (SA)a Monte Carlo based algorithm for combinatorial optimization problems inspired by statistical mechanics in thermodynamics with the statistical ensemble of the probability distribution over all possible states of a system described by a Markov chain, where its stationary distribution converts to an optimal distribution during a cooling process after reaching the equilibrium. Thus, the annealing algorithm simulates a nonstationary finite state Markov chain whose state space is the domain of the cost function called energy to be minimized

^{[1]}.^{[2]}## Table of Contents

## History

The annealing algorithm is an adaptation of the Metropolis–Hastings algorithm to generate sample states of a thermodynamic system, invented by Marshall Rosenbluth and published by Nicholas Metropolis et al. in 1953^{[3]}^{[4]}, later generalized by W. Keith Hastings at University of Toronto^{[5]}. According to Roy Glauber and Emilio Segrè, the original algorithm was invented by Enrico Fermi and reinvented by Stanislaw Ulam^{[6]}.SA was independently described by Scott Kirkpatrick, C. Daniel Gelatt and Mario P. Vecchi in 1983

^{[7]}, at that time affiliated with IBM Thomas J. Watson Research Center, Yorktown Heights, and by Vlado Černý from Comenius University, Bratislava in 1985^{[8]}.## Quotes

In the 2003 conference proceedingsCelebrating the 50th Anniversary of the Metropolis Algorithm^{[9]}^{[10]}, Marshall Rosenbluth describes the algorithm in the following beautifully concise and clear manner^{[11]}:A simple way to do this [sampling configurations with the Boltzmann weight], as emerged after discussions with Teller, would be to make a trial move: if it decreased the energy of the system, allow it; if it increased the energy, allow it with probability exp(−ΔE/kT) as determined by a comparison with a random number. Each step, after an initial annealing period, is counted as a member of the ensemble, and the appropriate ensemble average of any quantity determined.## Applications

SA has multiple applications in discrete NP-hard optimization problems such as the Travelling salesman problem, in machine learning, in training of neural networks, and in the domain of computer games and computer chess in automated tuning as elaborated by Peter Mysliwietz in his Ph.D. thesis^{[12]}to optimize the evaluation weight vector in Zugzwang. In its variant of temporal difference learning to adjust pattern weights in Morph, Robert Levinson at al. used simulated annealing as metaheuristic to set its own learning rate for each pattern, the more frequently a pattern is updated, the slower becomes its learning rate^{[13]}^{[14]}^{[15]}.## Algorithm

## Description

The control flow of the algorithm is determined by two nested loops, the outer loop over decreasing temperature simulates the cooling, and an inner loop times n Monte Carlo iterations. Each time a randomly picked neighbor state inside the inner loop provides a better energy or fitness than the current state, the neighbor becomes the new current and even new optimum if fitter than fittest so far. Otherwise, if the neighbor fitness does not exceed current, it might still become current depending on the positive fitness or energy difference ΔE, and absolute temperature T, with a probability p according to the Boltzmann factor:where k the Boltzmann constant, and e base of the exponential function whose negative exponent ensures the [0, 1] probability interval. Accepting worse solutions is a primary feature of SA, and important to stop greedy exploitation a local optimum but to explore other areas - higher temperatures favor exploration, while decreasing temperatures make the algorithm to behave greedier in favoring exploitation of the hopefully global optimum.

## Animation

^{[16]}With the high temprature, the numerous local maxima are left quickly through the strong noise movement -

but the global maximum is reliably found because of cooling temperature is no longer sufficient to leave it.

## Pseudo Code

The C like pseudo code is based on Peter Mysliwietz' description as given in his Ph.D. thesis^{[17]}. Several neighbor functions used to modify the weight vector were tried, where one randomly chosen element changed randomly performed well. The fitness function inside the inner loop is of course the most time consuming part. For Zugzwang, Mysliwietz used a database of 500 test-positions with a search depth of one ply, which took about three minutes on a T 800 Transputer per iteration - the higher the hit rate of found expert moves, the fitter. The whole optimization used a tHight to tLow ratio of 100, a reduction factor r of 0.95, and n=40 inner iterations.^{[18]}^{[19]}## See also

## Selected Publications

^{[20]}## 1948 ...

1948).Note on census-taking in Monte Carlo calculations. Los Alamos National Laboratory, pdf1949).The Monte Carlo Method. Journal of the American Statistical Association, Vol. 44, No. 247, pdf## 1950 ...

1953).Equation of State Calculations by Fast Computing Machines. Journal of Chemical Physics, Vol. 21, No. 61954).Further Results on Monte Carlo Equations of State. Journal of Chemical Physics, Vol. 22, No. 5, pdf## 1970 ...

1970).Monte Carlo Sampling Methods Using Markov Chains and Their Applications. University of Toronto, Biometrika, Vol. 57, No. 1, pdf## 1980 ...

1983).Optimization by Simulated Annealing. Science, Vol. 220, No. 4598, pdf1985).Thermodynamical approach to the traveling salesman problem: An efficient simulation algorithm. Journal of Optimization Theory and Applications, Vol. 45, No. 11985).Analysis of simulated annealing for optimization. 24th IEEE Conference on Decision and Control1987).Analysis of Simulated Annealing Type Algorithms. Ph.D. thesis, MIT, advisor: Sanjoy K. Mitter, pdf1987).The Beginning of the Monte Carlo Method. Los Alamos Science Special, pdf1989).Simulated Annealing Algorithms. IEEE Circuits and Devices Magazine, pdf## 1990 ...

1990).Threshold Accepting: A General Purpose Optimization Algorithm Appearing Superior to Simulated Annealing. Journal of Computational Physics, Vol. 90, No. 1^{[21]}1991).On the convergence of “Threshold Accepting”. Applied Mathematics and Optimization, Vol. 24, No. 11991).Artificial Intelligence or Stochastic Relaxation: Simulated Annealing Challenge. Heuristic Programming in AI 21993).Monte Carlo Go. pdf1993).Applied Simulated Annealing. Lecture Notes in Economics and Mathematical Systems, Vol. 396, Springer1994).Experience-Based Creativity. Artificial Intelligence and Creativity: An Interdisciplinary Approach, Kluwer1994).Konstruktion und Optimierung von Bewertungsfunktionen beim Schach.Ph.D. thesis (German)1999).Stochastic Optimization: Stochastic Approximation and Simulated Annealing. in John G. Webster (ed.) (1999). Encyclopedia of Electrical and Electronics Engineering, Vol. 20, John Wiley & Sons, pdf## 2000 ...

2002).Learning a Game Strategy Using Pattern-Weights and Self-play. CG 2002, pdf2002).Empirical study of Simulated Annealing aimed at improved multiobjective optimization. pdf2003).The Monte Carlo Method in Physical Sciences: Celebrating the 50th Anniversary of the Metropolis Algorithm. AIP Conference Proceedings^{[22]}2004).The MORPH Project in 2004. ICGA Journal, Vol. 27, No. 42005).Marshall Rosenbluth and the Metropolis Algorithm. Physics of Plasmas, Vol. 12, No. 5, pdf## 2010 ...

2010).Implementing a Parallel Simulated Annealing Algorithm. Lecture Notes in Computer Science, Volume 6067, Springer2011).Co-operative, Parallel Simulated Annealing for the VRPTW. Computational Collective Intelligence. Technologies and Applications, Lecture Notes in Computer Science, Springer^{[23]}2011).Simulated Annealing. in Berthold Vöcking et al. (eds.) (2011).Algorithms Unplugged. Springer2013).Combining Simulated Annealing and Monte Carlo Tree Search for Expression Simplification. CoRR abs/1312.0841## Forum Posts

## External Links

## Simulated Annealing

## Related Topics

## Misc

^{[24]}, YouTube Video## References

1985).Analysis of simulated annealing for optimization. 24th IEEE Conference on Decision and Control1953).Equation of State Calculations by Fast Computing Machines. Journal of Chemical Physics, Vol. 21, No. 61987).The Beginning of the Monte Carlo Method. Los Alamos Science Special, pdf1970).Monte Carlo Sampling Methods Using Markov Chains and Their Applications. University of Toronto, Biometrika, Vol. 57, No. 1, pdf1983).Optimization by Simulated Annealing. Science, Vol. 220, No. 4598, pdf1985).Thermodynamical approach to the traveling salesman problem: An efficient simulation algorithm. Journal of Optimization Theory and Applications, Vol. 45, No. 12003).The Monte Carlo Method in Physical Sciences: Celebrating the 50th Anniversary of the Metropolis Algorithm. AIP Conference Proceedings2005).Marshall Rosenbluth and the Metropolis Algorithm. Physics of Plasmas, Vol. 12, No. 5, pdf1994).Konstruktion und Optimierung von Bewertungsfunktionen beim Schach.Ph.D. thesis (German)1994).Experience-Based Creativity. Artificial Intelligence and Creativity: An Interdisciplinary Approach, Kluwer2002).Learning a Game Strategy Using Pattern-Weights and Self-play. CG 2002, pdf2000).Machine Learning in Games: A Survey. Austrian Research Institute for Artificial Intelligence, OEFAI-TR-2000-3, pdf1994).Konstruktion und Optimierung von Bewertungsfunktionen beim Schach.Ph.D. thesis, 7.4. Simulated Annealing, 7.4.2. Beschreibung des Algorithmus, Abb. 29, pp. 146 (German)## What links here?

Up one Level