Cross-entropy method explained
The cross-entropy (CE) method is a Monte Carlo method for importance sampling and optimization. It is applicable to both combinatorial and continuous problems, with either a static or noisy objective.
The method approximates the optimal importance sampling estimator by repeating two phases:[1]
- Draw a sample from a probability distribution.
- Minimize the cross-entropy between this distribution and a target distribution to produce a better sample in the next iteration.
Reuven Rubinstein developed the method in the context of rare-event simulation, where tiny probabilities must be estimated, for example in network reliability analysis, queueing models, or performance analysis of telecommunication systems. The method has also been applied to the traveling salesman, quadratic assignment, DNA sequence alignment, max-cut and buffer allocation problems.
Estimation via importance sampling
Consider the general problem of estimating the quantity
\ell=Eu[H(X)]=\intH(x)f(x;u)rm{d}x
,
where
is some
performance function and
is a member of some
parametric family of distributions. Using
importance sampling this quantity can be estimated as
,
where
is a random sample from
. For positive
, the theoretically
optimal importance sampling
density (PDF) is given by
.
This, however, depends on the unknown
. The CE method aims to approximate the optimal PDF by adaptively selecting members of the parametric family that are closest (in the
Kullback–Leibler sense) to the optimal PDF
.
Generic CE algorithm
- Choose initial parameter vector
; set t = 1.
- Generate a random sample
from
- Solve for
, where
}_ \frac \sum_^N H(\mathbf_i)\frac \log f(\mathbf_i;\mathbf)
- If convergence is reached then stop; otherwise, increase t by 1 and reiterate from step 2.
In several cases, the solution to step 3 can be found analytically. Situations in which this occurs are
belongs to the
natural exponential family
is
discrete with finite
support
} and
, then
corresponds to the
maximum likelihood estimator based on those
.
Continuous optimization - example
The same CE algorithm can be used for optimization, rather than estimation. Suppose the problem is to maximize some function
, for example,
. To apply CE, one considers first the
associated stochastic problem of estimating
}(S(X)\geq\gamma)for a given
level
, and parametric family
\left\{f( ⋅ ;\boldsymbol{\theta})\right\}
, for example the 1-dimensional
Gaussian distribution,parameterized by its mean
and variance
(so
\boldsymbol{\theta}=(\mu,\sigma2)
here).Hence, for a given
, the goal is to find
so that
DKL(rm{I}\{S(x)\geq\gamma\
}\|f_)is minimized. This is done by solving the sample version (stochastic counterpart) of the KL divergence minimization problem, as in step 3 above.It turns out that parameters that minimize the stochastic counterpart for this choice of target distribution andparametric family are the sample mean and sample variance corresponding to the
elite samples, which are those samples that have objective function value
.The worst of the elite samples is then used as the level parameter for the next iteration.This yields the following randomized algorithm that happens to coincide with the so-called Estimation of Multivariate Normal Algorithm (EMNA), an
estimation of distribution algorithm.
Pseudocode
// Initialize parameters μ := −6 σ2 := 100 t := 0 maxits := 100 N := 100 Ne := 10 // While maxits not exceeded and not converged while t < maxits and σ2 > ε do // Obtain N samples from current sampling distribution X := SampleGaussian(μ, σ2, N) // Evaluate objective function at sampled points S := exp(−(X − 2) ^ 2) + 0.8 exp(−(X + 2) ^ 2) // Sort X by objective function values in descending order X := sort(X, S) // Update parameters of sampling distribution via elite samples μ := mean(X(1:Ne)) σ2 := var(X(1:Ne)) t := t + 1 // Return mean of final sampling distribution as solution return μ
Related methods
See also
Journal papers
- De Boer, P.-T., Kroese, D.P., Mannor, S. and Rubinstein, R.Y. (2005). A Tutorial on the Cross-Entropy Method. Annals of Operations Research, 134 (1), 19–67.http://www.maths.uq.edu.au/~kroese/ps/aortut.pdf
- Rubinstein, R.Y. (1997). Optimization of Computer Simulation Models with Rare Events, European Journal of Operational Research, 99, 89–112.
Software implementations
Notes and References
- Rubinstein, R.Y. and Kroese, D.P. (2004), The Cross-Entropy Method: A Unified Approach to Combinatorial Optimization, Monte-Carlo Simulation, and Machine Learning, Springer-Verlag, New York .