EDAs remove the operators of crossover and mutation. In each generation a subset of the population is selected and the distribution of the individuals of this subset is estimated. The individuals of the population for the next generation are obtained sampling the estimated distribution. Although any selection method could be applied, the most common one is the selection of the best individuals of the population.
The first EDAs were developed for discrete spaces. Later, they were
adapted to continuous domains. We can distinguish two types of EDAs,
whether they take into account dependencies between the variables or
not. One of the most used among the EDAs that do not consider
dependencies is (Univariate Marginal Distribution Algorithm
for continuous domains) [LELP00]. In every generation and
for every variable the
carries out some statistical test in
order to find the density function that best fits the variable. Once
the densities have been identified, the estimation of parameters is
performed by their maximum likelihood estimates. If all the
distributions are normal, the two parameters are the mean and the
standard deviation. This particular case will be denoted
(Univariate Marginal Distribution Algorithm for Gaussian models).
Among the other type of EDAs, we can consider
(Estimation
of Gaussian Network Algorithm) [LELP00] whose good
results in function optimization are reported by Bengoetxea and Miqu�lez
[BM02]. In each generation,
learns the
Gaussian network structure by using a Bayesian score that gives the
same value for Gaussian networks reflecting the same conditional
dependencies are used. Next, it calculates estimations for the
parameters of the Gaussian network structure.
In the experiments we have used the parameters reported by Bengoetxea and T. Miqu�lez [BM02]: a population of 2000 individuals, initialized using a uniform distribution, from which a subset of the best 1000 individuals are selected to estimate the density function, and the elitist approach was chosen (the best individual is included for the next population and 1999 individuals are simulated). Each algorithm has been run 30 times with a stop criterion of 300,000 evaluations of the fitness function.
The results of EDAs are compared with the results of a RCGA with CIXL2
of parameters and
. We performed an ANOVA I
analysis where the three levels of the factor are the different
algorithms: RCGA with CIXL2,
and
. We also
carried out a multiple comparison test.
Table 4 shows the average values and standard
deviations for 30 runs for each algorithm. Table 11
in Appendix A shows how, for all the functions
excepting , the type of algorithm has a significant effect
over the linear model and exist inequality of the variances of the
results (Levene test). So, we have used Tamhane test for all the
functions and Bonferroni test for
. Table
17 (Appendix A) shows the
results of the multiple comparison test and the ranking established by
the test.
For the results are very similar. The fitness behaves as a
singular random variable with sample variance near 0 and the
statistical tests are not feasible.
For the results of CIXL2 are significantly better than the
results of
and
. The same situation occurs for
,
,
and
, with the exception that
in these four functions there are no significant differences between
the two EDAs. For
,
and
achieve the
best results, significantly better than CIXL2. For
,
is significantly better than
and CIXL2, but there are no
differences between these two. For
, CIXL2 obtains the best
results, but there are no significant differences among the three
algorithms.
The estimation of the distribution function of the best individuals of
the population performed by EDAs is an advantage in ,
unimodal and separable, and
and
whose optima are
regularly distributed. The results of EDAs for
are better
than the results of CIXL2, but the results for
are
worse. The results for
of all the algorithms are very
similar. For non-separable unimodal functions, such as
and
, the interdependence among their variables should favor the
performance of
over
and CIXL2. Nevertheless,
CIXL2 achieves the best results for these two functions. For
multimodal separable functions,
and
, it is
difficult to identify the distribution of the best individuals and the
performance of EDAs is below the performance of CIXL2.
For extremely complex functions, such as and
, the
results are less conclusive. For
the best results are
obtained with
, and there are no differences between
and CIXL2. For
, CIXL2 achieves the best
results, but the differences among the three algorithms are not
statistically significant.
Domingo 2005-07-11