Professional Documents
Culture Documents
Institute of Geophysics, Polish Academy of Sciences, Ks. Janusza 64, 01-452 Warsaw, Poland
a r t i c l e i n f o
Article history:
Received 14 November 2012
Received in revised form 18 March 2013
Accepted 31 March 2013
Available online 10 April 2013
Keywords:
Differential Evolution
Evolutionary algorithm
Memetic algorithm
Memetic computing
Self-adaptation
Global numerical optimization
a b s t r a c t
Differential Evolution (DE) is one of the most popular optimization methods for real-valued
problems and a large number of its variants have been proposed so far. However, bringing
together different ideas that already led to successful DE versions is rare in the literature. In
the present paper a novel DE algorithm is proposed, in which three among the most ef-
cient concepts already applied separately within DE framework are gathered together,
namely: (1) the adaptation of algorithm control parameters and probabilities of using dif-
ferent mutation strategies; (2) the use of NelderMead algorithm as a local search method
hybridized with DE; and (3) the splitting mutation into Global and Local models, when
Local mutation model is based on the concept of neighborhood of individuals organized
on a ring topology. The performance of the novel algorithm, called Adaptive Memetic DE
with Global and Local neighborhood-based mutation operators is compared with 13 differ-
ent DE variants on a set of 25 popular problems which include rotated, shifted and hybrid
composition functions. It is found that, although none DE algorithm outperforms all the
others for the majority of problems, on average the proposed approach perform better than
all 13 DE algorithms selected for comparison.
The proposed algorithm is another heuristic approach developed to solve optimization
problems. The question may arise, whether proposing novel methods is useful as No Free
Lunch theorems for optimization state that the expected performance of all possible heu-
ristics on all possible problems is equal. In the last section of the paper the limitations and
implications of No Free Lunch theorems are discussed based on rich, but unfortunately fre-
quently neglected literature. A very simple continuous and differentiable minimization
problem is proposed, for which it is empirically veried that each among considered 14
DE algorithms perform poorer than random sampling. It is also empirically shown that
when all considered DE algorithms search for the maximum of the proposed problem, they
found lower minimum than DE algorithms searching for the minimum. Such result is not
unexpected according to No Free Lunch theorems and should be considered as a precaution
from generalization of good performance of heuristic optimization methods.
2013 Elsevier Inc. All rights reserved.
1. Introduction
Differential Evolution (DE) [123,124,113] is one of the most popular Evolutionary Algorithms (EAs) for solving continuous
single-objective optimization problems, due to its simplicity and reasonable performance. Without the loss of generality, the
minimization problems are considered in this study. The goal is to nd the global optimum vector x
such that
0020-0255/$ - see front matter 2013 Elsevier Inc. All rights reserved.
http://dx.doi.org/10.1016/j.ins.2013.03.060
5
k=1
S
k;g
(23)
hence contrary to SADE, in AM-DEGL the probability of using each strategy cannot be reduced to lower value than 0.01. In
fact it must be slightly higher as, following SADE, 0.01 constant is also used in Eq. (22). The larger minimum chance of being
used than used in SADE is advantageous to avoid a situation when a good strategy is unused for a number of generations due
to occasional poor performance, which may happen by chance as the adaptation is based on the results from the single gen-
eration only. The variant of AM-DEGL which uses equations proposed in SADE (Eq. (14) in [115] with LP = 1), instead of Eqs.
(22) and (23), is tested in Section 5.4.
4.5. Local search algorithm
Various local search methods are used by Memetic DE algorithms, from a very simple random walk [78] or chaotic search
[69] to heuristics like Adaptive hill descending with simplex crossover [98], Simulated Annealing, Hooke Jeeves Algorithm
[132], Rosenbrock method or NelderMead simplex [18]. Since the publication of [102] meta-Lamarckian learning (i.e.
hybridizing global search heuristic with more than one local search method) with adaptive probability of commencing each
local algorithm was also tested within DE framework [132,18]. However, in both mentioned papers only simple DE variants,
without adaptation of control parameters and with a single mutation strategy were used. Because in AM-DEGL a number of
Global and Local mutation strategies are applied, simple Lamarckian learning is chosen, namely a single NMA algorithm with
pre-dened probability to be commenced is adopted as local search method. Hybridization of DE with single local search
algorithm is also frequent in the literature [98,78,69]. NMA is chosen due to its ability to leave the basin of attraction, sig-
nicant, long-lasting popularity and a vast number of successful applications.
The key point is which individuals should be allowed to carry local search. Among Memetic DE algorithms various ap-
proaches were tested. In [98,69] and in the case of RA used in [18], only the best individuals have a chance to perform local
search. This saves the function calls, but allows speeding up the exploration of only the one local optimum that seems the
most promising at the moment. Contrary, in [78,132] and in the case of NMA used in [18], any individual may followthe local
search. Such approach may allow recognizing the presence of promising local optima in the vicinity of various individuals,
but frequently waste much more function calls.
In AM-DEGL an intermediate approach is proposed, which benet from the idea of Local mutation model only individ-
uals that are better than all other individuals within p% neighborhood have a chance to follow the local search (in AM-DEGL
p% is set to 20%, see discussion in Section 4.2). This way the number of individuals that have a chance to commence local
search may vary from 0 to almost 10% of the population, depending on relation in tness among neighboring individuals.
In practice it is observed that for 30D problems, when NP = 5D, usually 45 individuals in a generation are better than all
others within their neighborhood, and only they have a chance to follow local search. Because the information exchange be-
tween this few individuals is slow, the probability that they are located within the basins of attractions of different local min-
ima is higher than for neighboring individuals.
Another very important issue is the probability that an individual that is allowed to follow local search would in fact com-
mence it, and the length of local search. In the literature two options prevail (1) frequent but short or (2) rare but longer
local search. The rst option performs better for problems with simpler tness landscape when greedy local search meth-
ods are used, the second is advantageous if tness landscape is more complicated and more advanced heuristics that may in
fact leave the basin of attraction are used.
The rst option is applied in [98,78,69]. The local search may be commenced at every generation for the best individual
only [98] or even for every individual in the population [78]. In [78] maximum number of function calls available for each
local search is set to only 10. In [69] chaotic local search may last up to D/5 function calls, but terminates earlier when for the
174 A.P. Piotrowski / Information Sciences 241 (2013) 164194
rst time the better solution than the parents one is found. Interestingly, in [98] the contrary approach is used and the local
search is carried on as long as each offspring generated by simplex crossover is better than the previous one.
The second option is used in [18,132] and seems to be a good choice when NMA is used, because NMA is able to leave the
basin of attraction of local optimum. But this approach is more greedy in terms of function calls as NMA requires moving
D + 1 points throughout the search space.
Based on the above discussion, AM-DEGL follow the second option rare but longer local search. In AM-DEGL the NMA
may be employed for each individual which is better than all other individuals within p% neighborhood with the probability
LSp = 0.005, and last for 100D function calls.
NMA requires D + 1 points to be initialized. In AM-DEGL the rst point is represented by the individual for which the local
search is commenced which is the best one within own neighborhood (dened according to a ring topology). A concept
how to select additional D points is needed. The chosen points should show some similarities (if they span the whole search
space, the local search would probably be unsuccessful), but should also introduce some diversity to the simplex (to exploit
exploration capabilities of NMA and avoid problems with so called lost dimension). Hence in AM-DEGL the selection is split
into halves. The rst D/2 individuals are neighbors of x
i,g
with lower indexes i (i.e. x
i1,g
, . . . , x
iD/2,g
). These individuals are
within p% = 20% neighborhood of x
i,g
hence they exchange information among themselves and with x
i,g
frequently by means
of Local mutation model. This may be advantageous or not, depending on how similar closest neighbors are using too
similar individuals as starting points may diminish NMA search capabilities. Hence, as the another D/2 points the
consecutive individuals that have higher indexes than x
i,g
but are not within p% = 20% neighborhood of x
i,g
are used (i.e.
x
i+D/2+1,g
, . . . ,x
i+D,g
). They are localized close to x
i,g
according to a ring topology (hence may also show some similarities with
x
i,g
), however they exchange information by means of Local mutation model directly among themselves, but not with the
other points to be used within NMA (i.e. x
i,g
, . . . , x
iD/2,g
). One should remember that the NP individuals are organized on
a closed ring, hence the next index after NP is 1. If the problem is multimodal, one may expect that within such formed sim-
plex a number of local minima may exists, and the goal of NMA is to locate and exploit the best of them, what is time con-
suming. This explains why the proposed length of the local search is longer than in [18], where each local search lasted 500
function calls and mostly 10-dimensional benchmark problems were considered. A variant of AM-DEGL which uses the clos-
est individuals to x
i
in the search space according to Euclidean norm as additional D points is tested in Section 5.4.
In AM-DEGL the worst individual in DE population x
worst,g
is replaced by the best point found by the NMA. Hence the
local search helps improving the performance of the poorest part of the population. Note that x
worst,g
is always poorer
than the best solution found by NMA that start from x
i,g
. In the implementation used in the present paper (see pseudo-
code, Fig. 1) x
worst,g
is replaced immediately, but if parallel computing is used it may be replaced when the NP loop
terminates.
5. Results and discussion
5.1. Benchmark problems
The performance of AM-DEGL is tested on 10-, 30- and 50-dimensional versions of standard 25 benchmark problems pre-
pared for CEC 2005 [126]. Each algorithm is run 30-times on each version of every test problem. Since the publication of
[119] it is known that EAs should be tested on rotated problems, as many EAs lack rotational invariance properties. Also most
DE algorithms are not rotationally invariant due to the crossover operation, hence they performbetter for separable and non-
rotated problems. The CEC 2005 set of problems is chosen as it includes rotated, shifted and composition functions. Very
brief description of each problem is provided in Table 1, for more details the reader is referred to [126].
5.2. Algorithms for comparison
In the present paper the performance of the proposed AM-DEGL is compared with 13 different DE algorithms. The pop-
ulation size NP = 5D is set for AM-DEGL and most other DE methods, as this value lies in the middle of most frequently sug-
gested population sizes found in the literature [124,48,113,80,98,142,156]. However, due to their specic capabilities, some
Composite or Memetic DE algorithms require small population size, hence they are applied with NP = 1D; on the contrary,
DE-SG require NP = 10D (as its population is split into D groups with 10 individuals each [109]). The values of control param-
eters suggested in the source papers are used, with the exception of DE-SG, which was designed to solve problems when
large number of function calls is available in this case to quicken convergence two control parameters were modied,
see point 2 below. Some authors proposed more than one version of their algorithm in the source papers, in such case
the best performing variant is chosen and dened below. The algorithms for comparison include:
1. DE original version of Differential Evolution [123,124,113] with DE/rand/1/bin; NP = 5D; F = 0.8; CR = 0.5;
2. DE-SG DE with Separated Groups [109,110]; NP = 10D; PNI = 10; MigProb = 1/PNI;
3. DEGL DE with Global and Local neighborhood-based mutation operators [29]; NP = 5D; self-adaptive weight factor is
used;
4. SADE Self-Adaptive DE [115]; NP = 5D;
A.P. Piotrowski / Information Sciences 241 (2013) 164194 175
5. SspDE Self-Adaptive DE [105]; NP = 5D;
6. JADE [156]; NP = 5D; with an archive;
7. AdapSS-JADE Adaptive Strategy Selection-based JADE [55]; NP = 5D; with Averaged Normalized Reward method;
8. DECLS DE algorithm based on chaotic local search [69]; NP = 1D;
9. DEahcSPX DE with adaptive hill-climbing (in case of minimization descending) crossover-based local search [98];
NP = 1D;
10. CoDE Composite DE [139]; NP = 1D; note that in this approach three offspring are generated for each parent and t-
ness function is called for each of them, hence lower population size is required;
11. EPSDE Ensemble of mutation strategies and parameters values for DE [85]; NP = 5D;
12. SFMDE Super-t Memetic DE [18]; NP = 5D;
13. CDE Clustering-based DE [17]; NP = 5D.
5.3. Comparison
Tables 24 show the average best tness obtained by each algorithm for every 10-, 30- and 50-dimensional problem,
respectively, accompanied by the standard deviation. Three statistical procedures are employed to test the differences be-
tween the results obtained by different algorithms.
First, the version of Friedman test [47] extended by Iman and Davenport (see [35,49]) is applied to test the hypothesis
that there is no signicant difference between results obtained by means of all applied algorithms. Three F
F
statistics:
F
F
= 16.58, F
F
= 12.15 and F
F
= 20.74, that were obtained for 10-, 30- and 50-dimensional problems, respectively, show that
the hypothesis that the performance of all methods is similar may be rejected at very high level of signicance (three cor-
responding p-values are below 10
7
) and another tests may be applied to detect the best methods.
Than, a statistical signicance of difference between the results obtained by means of AM-DEGL and other algorithms is
tested by means of rank-based Holm variant [60] of one-tailed Bonferroni-Dunn test [35,49] that adjusts p-values when a
number of inter-comparisons are made. Holm procedure compares aggregated over all 25 problems performance of all
algorithms against the control method (AM-DEGL). The adjusted p-values (APv) and information whether AM-DEGL is sig-
nicantly better than each among 13 competitive algorithms at 5% signicance level are reported in the second lowest row in
Tables 24 (denoted as H/APv). Note that ++ means that AM-DEGL is signicantly better than the competitive algorithm, ==
means no signicant difference (if the APv value is higher than 0.5, what happens a few times for 10-dimensional problems,
the Holm procedure with competitive algorithm as a control method is applied to test if the competitive algorithm is signif-
icantly better than AM-DEGL in such cases tests always showed that the difference is not signicant).
Finally, for each algorithm (not only AM-DEGL) the statistical signicance of its performance against the performance of
13 other algorithms is assessed for each problem separately by very popular (although sometimes criticized when applied
for multi-comparisons, see [49]) Wilcoxon rank-sum test [148] at 5% signicance level (see +/=/ columns in Tables 24).
Table 1
Summary of CEC 2005 test problems.
Problem Type f (x
1
9
4
1
7
7
Table 2 (continued)
AM-DEGL DE DE-SG DEGL SADE
mean Std +/=/ Mean Std +/=/ Mean Std +/=/ Mean Std +/=/ Mean Std +/=/
F15 215.076 1.59E+2 7/5/1 255.724 1.70E+2 5/5/3 244.643 1.73E+2 6/6/1 342.720 1.54E+2 1/3/9 466.038 1.43E+2 0/1/12
F16 216.704 4.74E+0 10/3/0 217.055 6.30E+0 10/3/0 216.856 5.12E+0 10/3/0 244.161 2.18E+1 1/3/9 246.752 2.00E+1 1/3/9
F17 224.472 6.65E+0 7/5/1 225.767 6.11E+0 7/5/1 221.583 5.14E+0 11/2/0 248.632 1.69E+1 2/2/9 248.799 2.05E+1 2/2/9
F18 610.000 2.49E+2 3/7/3 674.684 2.65E+2 2/7/4 739.543 2.22E+2 1/5/7 861.838 1.83E+2 0/0/13 818.845 1.68E+2 1/3/9
F19 586.033 2.52E+2 4/7/2 643.059 2.60E+2 4/6/3 768.109 2.13E+2 1/3/9 884.725 1.65E+2 0/1/12 774.892 2.43E+2 0/4/9
F20 685.352 2.22E+2 4/4/5 721.275 2.34E+2 2/6/5 704.132 2.45E+2 2/6/5 931.768 8.66E+1 0/0/13 790.109 1.86E+2 1/4/8
F21 943.333 1.91E+2 4/6/3 986.674 2.20E+2 4/5/4 982.101 2.20E+2 4/5/4 1234.758 2.62E+2 0/2/11 1232.715 2.78E+2 0/2/11
F22 1121.949 1.77E+1 5/2/6 1117.379 2.19E+1 6/6/1 1111.589 1.73E+1 6/5/2 1143.897 1.10E+2 0/5/8 1159.528 4.64E+1 0/2/11
F23 1080.218 1.92E+2 4/7/2 1128.118 1.75E+2 2/6/5 1200.437 2.05E+2 1/3/9 1296.338 2.73E+2 0/3/10 1322.804 2.35E+2 0/1/12
F24 470.000 5.48E+1 2/11/0 510.000 1.14E+2 1/7/5 500.000 1.04E+2 1/7/5 530.258 1.30E+2 0/1/12 523.481 1.59E+2 1/7/5
F25 635.583 3.40E+0 7/6/0 644.110 3.17E+1 5/5/3 644.447 3.16E+1 5/5/3 853.563 3.57E+2 0/1/12 683.080 1.44E+2 1/3/9
H/APv 0.3747 (==) 0.2527 (==) 0.1750 (==) 1.55E9 (++) 3.54E10 (++)
W/Pav 4.80/1.96 4.24/2.28 4.36/2.64 1.24/9.84 1.00/8.44
CoDE EPSDE SFMDE CDE
F1 450.000 1.39E7 1/12/0 450.000 1.37E7 1/12/0 449.761 1.31E+0 1/0/12 450.000 2.22E7 2/11/0
F2 450.000 1.17E7 1/12/0 450.000 1.57E7 1/12/0 450.000 4.53E4 1/12/0 450.000 1.68E7 1/12/0
F3 2869.529 6.44E+3 4/4/5 441.493 3.39E+1 10/0/3 2.12E+04 9.58E+4 2/3/8 6936.786 7.92E+3 3/3/7
F4 450.000 1.46E7 2/11/0 450.000 1.41E7 2/11/0 406.592 2.00E+2 0/9/4 450.000 1.62E7 3/10/0
F5 310.000 1.03E7 1/12/0 310.000 5.84E8 1/12/0 229.331 3.70E+2 0/11/2 310.000 8.72E8 1/12/0
F6 390.399 1.22E+0 3/6/4 390.133 7.28E1 9/2/2 390.930 1.71E+0 3/6/4 392.868 1.26E+0 1/0/12
F7 179.921 3.83E2 5/1/7 179.878 7.24E2 4/0/9 179.410 7.92E1 1/1/11 179.973 2.96E2 8/5/0
F8 119.684 1.05E1 4/7/2 119.651 5.08E2 2/7/4 119.962 4.52E2 12/0/1 119.767 1.71E1 9/2/2
F9 329.735 5.18E1 4/8/1 330.000 1.39E7 4/9/0 326.086 1.67E+0 1/2/10 330.000 2.05E7 5/8/0
F10 318.492 5.94E+0 3/2/8 316.864 3.06E+0 1/1/11 320.803 5.55E+0 4/3/6 324.826 2.82E+0 8/3/2
F11 94.319 2.10E+0 2/6/5 96.196 8.26E1 1/0/12 94.969 1.06E+0 2/4/7 90.936 1.09E+0 12/1/0
F12 372.249 1.98E+2 3/7/10 457.046 6.40E+0 11/2/0 236.118 5.08E+2 3/7/10 434.691 8.43E+1 5/6/2
F13 129.584 9.89E2 6/3/4 129.341 8.16E2 2/2/9 129.220 2.76E1 2/2/9 129.563 1.23E1 5/4/4
F14 296.921 3.46E1 4/1/8 296.738 1.96E1 2/1/10 296.681 2.47E1 2/1/10 297.383 3.48E1 9/3/1
F15 276.855 1.70E+2 4/4/5 266.709 1.84E+2 5/7/1 359.704 1.43E+2 1/2/10 333.379 1.71E+2 3/2/8
F16 233.090 8.69E+0 4/1/8 235.340 6.86E+0 1/4/8 246.523 2.73E+1 1/3/9 219.455 1.21E+1 6/4/3
F17 229.826 8.53E+0 6/4/3 256.999 1.01E+1 1/1/11 250.679 2.35E+1 1/3/9 224.373 1.08E+1 8/5/0
F18 634.968 2.53E+2 5/6/2 616.280 2.55E+2 5/6/2 671.417 2.32E+2 1/5/7 493.333 2.45E+2 10/3/0
F19 645.123 2.56E+2 4/7/2 512.282 2.52E+2 7/6/0 705.301 2.17E+2 1/3/9 516.627 2.53E+2 5/8/0
F20 606.876 2.66E+2 7/5/1 611.842 2.51E+2 7/5/1 698.316 2.19E+2 2/4/7 557.336 2.53E+2 7/5/1
F21 924.882 2.17E+2 4/9/0 866.667 1.66E+2 6/7/0 1099.838 2.49E+2 2/1/10 931.815 1.61E+2 4/7/1
F22 1107.818 8.79E+1 6/5/2 1091.634 1.17E+2 7/4/1 1131.778 2.26E+1 1/5/7 1081.334 1.16E+2 8/5/0
F23 981.053 1.17E+2 5/6/2 1020.324 1.55E+2 7/4/2 1055.451 2.34E+2 3/7/3 1077.018 1.65E+2 4/6/3
F24 510.000 1.14E+2 1/7/5 480.000 7.61E+1 2/11/0 460.005 2.98E2 1/7/5 460.000 0.00E+0 7/6/0
F25 637.963 3.84E+0 5/3/5 637.616 2.33E+1 8/3/2 647.718 9.88E+0 1/3/9 634.358 2.70E+0 9/4/0
H/APv 0.1273 (==) 0.1613 (==) 6.20E6 (++) 1 (==)
W/Pav 3.76/3.28 4.28/3.56 1.96/6.88 5.72/1.88
1
7
8
A
.
P
.
P
i
o
t
r
o
w
s
k
i
/
I
n
f
o
r
m
a
t
i
o
n
S
c
i
e
n
c
e
s
2
4
1
(
2
0
1
3
)
1
6
4
1
9
4
Table 3
The mean tness values (mean) and its standard deviations (std) for 30-dimensional problems. +/=/ shows then how many of 13 competing methods the particular algorithm is signicantly better/equal/
signicantly worse, according to Wilcoxon rank-sum test at 5% signicance level. W/Pav how many times, averaged over all problems, the particular algorithm is signicantly better/signicantly worse than other
algorithms according to Wilcoxon test. H/APv adjusted p-values according to Holm test procedure with AM-DEGL as a control method; ++ means that AM-DEGL is signicantly better than the competitive algorithm,
== means that there is no signicant difference between AM-DEGL and competitive method at 5% signicance level, according to Holms procedure.
AM-DEGL DE DE-SG DEGL SADE
Mean Std +/=/ Mean Std +/=/ Mean Std +/=/ Mean Std +/=/ Mean Std +/=/
F1 450.000 3.47E7 2/11/0 449.939 1.24E2 0/0/13 450.000 6.92E8 2/11/0 450.000 1.21E7 2/11/0 450.000 1.13E7 2/11/0
F2 450.000 2.37E7 3/10/0 1.72E+4 2.70E+3 0/0/13 449.962 3.07E2 2/0/11 450.000 5.41E8 3/10/0 450.000 3.35E8 3/10/0
F3 1780.313 4.43E+3 13/0/0 1.34E+8 2.68E+7 0/0/13 7.23E+4 4.64E+4 10/1/2 4.26E+4 2.39E+4 11/1/1 8.21E+4 9.98E+4 10/2/1
F4 450.000 6.16E4 11/2/0 2.48E+4 3.38E+3 0/0/13 448.872 5.23E1 4/0/9 450.000 5.06E8 11/2/0 449.997 9.62E3 9/1/3
F5 309.980 9.75E2 12/0/1 5456.407 3.90E+2 0/0/13 526.089 3.50E+2 7/0/6 309.999 1.56E3 13/0/0 68.838 5.61E+2 10/1/2
F6 391.595 5.44E+0 9/1/3 5797.662 1.56E+3 0/0/13 482.619 1.54E+1 1/1/11 391.063 1.79E+0 12/1/0 391.311 1.88E+0 10/1/2
F7 179.998 4.08E3 13/0/0 19.031 2.74E+1 0/0/13 179.255 2.45E1 2/0/11 179.987 1.16E2 4/6/3 179.988 7.94E3 5/4/4
F8 119.983 1.43E2 13/0/0 119.044 4.42E2 0/11/2 119.045 3.92E2 0/11/2 119.063 5.88E2 0/11/2 119.051 4.18E2 0/11/2
F9 326.658 2.45E+0 7/0/6 208.680 8.20E+0 0/0/13 324.188 2.09E+0 6/0/7 296.559 2.20E+1 2/0/11 330.000 7.94E8 10/3/0
F10 297.067 1.06E+1 11/2/0 108.776 1.09E+1 0/0/13 209.218 9.66E+0 3/0/10 274.584 4.64E+1 6/4/3 278.311 9.01E+0 8/0/5
F11 113.845 3.95E+0 10/2/1 129.637 1.15E+0 0/0/13 121.330 1.16E+0 5/1/7 101.430 8.38E+0 13/0/0 113.543 6.33E+0 9/3/1
F12 1915.835 4.45E+3 8/5/0 3.89E+5 3.73E+4 0/0/13 1.52E+4 1.58E+4 4/1/8 7351.982 8.08E+3 5/2/6 1372.484 2.53E+3 7/6/0
F13 128.027 4.02E1 11/1/1 110.999 1.13E+0 0/0/13 126.232 2.89E1 5/1/7 123.932 4.13E+0 2/3/8 127.151 2.16E1 8/2/3
F14 287.285 4.67E1 7/3/3 286.542 1.72E1 0/0/13 287.005 1.80E1 5/1/7 287.520 3.86E1 10/3/0 287.175 2.16E1 7/2/4
F15 520.000 9.10E+1 1/8/4 492.964 4.38E+1 1/11/1 486.429 1.03E+2 6/4/3 461.621 1.11E+2 5/8/0 514.642 8.41E+1 1/10/2
F16 197.896 6.86E+1 8/4/1 366.187 1.60E+1 0/0/13 240.114 1.08E+1 4/3/6 272.790 1.38E+2 3/6/4 177.164 1.41E+1 9/4/0
F17 220.789 4.43E+1 8/5/0 389.383 1.59E+1 0/1/12 288.726 1.10E+1 3/1/9 272.092 1.44E+2 4/9/0 221.070 7.26E+1 8/5/0
F18 914.442 5.06E1 4/4/5 917.051 1.78E1 0/3/10 874.285 5.34E+1 3/10/0 912.946 1.95E+1 8/5/0 914.999 1.44E+0 3/5/5
F19 900.701 3.62E+1 9/3/1 917.008 1.72E1 0/4/9 895.479 4.35E+1 9/4/0 909.618 2.71E+1 7/5/1 915.141 1.41E+0 4/4/5
F20 914.877 1.17E+0 4/4/5 917.027 1.77E1 0/1/12 884.611 4.97E+1 4/9/0 912.698 1.95E+1 7/5/1 914.946 1.24E+0 4/4/5
F21 860.000 1.1E13 4/9/0 860.013 2.36E3 3/0/10 860.000 6.19E9 4/9/0 906.667 1.22E+2 0/4/9 860.000 2.3E13 4/9/0
F22 1230.939 1.88E+1 12/1/0 1318.997 7.78E+0 0/1/12 1262.048 1.19E+1 2/2/9 1252.624 1.86E+1 4/3/6 1243.142 1.79E+1 6/6/1
F23 894.164 1.58E4 6/7/0 894.168 1.44E3 4/1/8 894.165 2.57E4 6/7/0 922.360 1.02E+2 0/3/10 894.164 5.81E5 6/7/0
F24 460.000 0.00E+0 1/12/0 460.034 6.47E3 0/0/13 460.000 7.84E8 1/12/0 460.000 0.00E+0 1/12/0 460.000 0.00E+0 1/12/0
F25 469.656 6.06E1 13/0/0 565.429 2.08E+1 0/0/13 470.340 6.51E1 9/3/1 490.420 7.33E+1 1/2/10 470.326 6.28E1 9/3/1
H/APv 3.33E16 (++) 2.98E4 (++) 0.0214 (++) 0.0855 (==)
W/Pav 8.00/1.24 0.32/11.36 4.28/5.04 5.36/3.00 6.12/1.84
SspDE JADE AdapSS-JADE DECLS DEahcSPX
F1 450.000 6.31E8 2/11/0 450.000 8.17E8 2/11/0 450.000 8.88E8 2/11/0 450.000 5.81E8 2/11/0 450.000 5.15E8 2/11/0
F2 450.000 4.06E5 3/10/0 450.000 3.79E8 3/10/0 450.000 3.41E5 3/10/0 450.000 2.33E6 3/10/0 450.000 6.21E5 3/10/0
F3 1.85E+05 8.95E+4 7/2/4 7.58E+05 3.00E+5 3/1/9 7.45E+05 3.57E+5 3/1/9 2.26E+05 1.20E+5 6/3/4 3.74E+05 2.09E+5 4/2/7
F4 422.121 6.18E+1 2/1/10 449.990 3.40E2 8/1/4 449.978 3.36E2 7/0/6 432.075 2.51E+1 2/1/10 449.120 1.40E+0 5/1/7
F5 2144.586 4.34E+2 2/0/11 1236.255 3.18E+2 4/1/8 974.882 3.62E+2 6/0/7 1491.183 6.75E+2 3/2/8 200.769 1.29E+2 10/1/2
F6 402.045 2.03E+1 4/2/7 396.426 9.05E+0 5/2/6 402.020 1.84E+1 4/2/7 394.317 1.24E+1 8/2/3 391.496 1.80E+0 9/1/3
F7 179.981 1.18E2 4/0/9 179.992 5.32E3 4/7/2 179.990 5.47E3 5/4/4 179.988 1.18E2 4/5/4 179.994 8.09E3 8/3/2
F8 119.046 4.82E2 0/11/2 119.047 4.24E2 0/11/2 119.052 5.31E2 0/11/2 119.056 5.05E2 0/11/2 119.046 6.07E2 0/11/2
F9 330.000 7.12E8 10/3/0 277.907 5.74E+0 1/0/12 329.012 2.86E+0 8/1/4 329.503 7.27E1 8/1/4 311.035 5.57E+0 5/0/8
F10 301.378 4.58E+0 11/2/0 288.959 6.89E+0 8/3/2 286.469 8.01E+0 8/2/3 291.466 9.46E+0 8/3/2 268.651 5.94E+1 6/4/3
F11 116.984 1.26E+0 9/1/3 118.628 1.65E+0 7/1/5 118.910 1.50E+0 7/1/5 115.171 3.04E+0 10/2/1 124.541 8.47E+0 2/0/11
F12 2204.386 3.30E+3 7/6/0 2099.096 3.05E+3 7/6/0 3935.267 5.42E+3 6/6/1 2579.561 4.53E+3 7/6/0 2140.800 2.91E+3 7/6/0
F13 127.986 1.85E1 11/1/1 124.816 4.32E1 4/0/9 125.189 5.75E1 4/1/8 128.782 1.47E1 13/0/0 127.283 6.08E1 8/2/3
F14 287.168 2.23E1 7/2/4 287.600 2.38E1 11/2/0 287.571 1.78E1 11/2/0 287.372 3.90E1 9/2/2 286.924 2.63E1 3/3/7
(continued on next page)
A
.
P
.
P
i
o
t
r
o
w
s
k
i
/
I
n
f
o
r
m
a
t
i
o
n
S
c
i
e
n
c
e
s
2
4
1
(
2
0
1
3
)
1
6
4
1
9
4
1
7
9
Table 3 (continued)
AM-DEGL DE DE-SG DEGL SADE
Mean Std +/=/ Mean Std +/=/ Mean Std +/=/ Mean Std +/=/ Mean Std +/=/
F15 513.400 7.41E+1 1/9/3 491.911 9.23E+1 1/11/1 476.667 9.35E+1 4/9/0 456.718 8.51E+1 10/3/0 507.154 8.60E+1 1/11/1
F16 183.381 6.74E+1 11/1/1 192.856 6.37E+1 7/5/1 213.502 1.01E+2 7/4/2 252.027 1.23E+2 4/1/8 204.940 8.18E+1 9/4/0
F17 212.292 2.27E+1 9/4/0 265.870 9.38E+1 5/4/4 260.366 9.59E+1 6/4/3 235.371 8.64E+1 8/5/0 279.492 8.45E+1 3/7/3
F18 909.721 2.72E+1 6/7/0 914.430 8.24E1 4/4/5 913.954 6.93E1 8/2/3 916.899 2.06E+0 0/4/9 914.084 9.62E1 8/2/3
F19 913.685 1.97E+1 7/4/2 914.388 6.94E1 5/3/5 913.990 5.00E1 7/2/4 916.640 2.39E+0 0/6/7 913.837 3.36E1 7/2/4
F20 910.564 2.74E+1 10/2/1 914.616 1.05E+0 4/4/5 914.042 6.56E1 7/2/4 916.729 2.26E+0 0/7/6 913.804 2.94E1 7/2/4
F21 860.000 4.4E13 3/10/0 860.000 2.2E13 2/11/0 860.000 2.3E13 3/10/0 860.000 3.2E13 2/11/0 860.000 1.4E13 4/9/0
F22 1247.129 1.83E+1 5/7/1 1239.633 1.74E+1 6/7/0 1245.344 1.23E+1 6/6/1 1242.046 1.80E+1 6/6/1 1247.500 1.62E+1 5/7/1
F23 894.164 9.23E5 5/8/0 907.592 7.35E+1 1/1/11 894.164 3.49E4 6/7/0 894.165 5.82E4 6/7/0 894.172 1.59E2 2/2/9
F24 460.000 5.7E13 1/12/0 460.000 0.00E+0 1/12/0 460.000 0.00E+0 1/12/0 460.000 0.00E+0 1/12/0 460.000 0.00E+0 1/12/0
F25 472.049 1.32E+0 3/4/6 471.705 6.39E1 3/4/6 471.758 5.64E1 3/4/6 472.893 2.75E+0 3/3/7 470.200 6.31E1 9/3/1
H/APv 0.0154 (++) 0.0159 (++) 0.0214 (++) 0.0083 (++) 0.0050 (++)
W/Pav 5.60/2.60 4.24/3.88 5.28/3.16 4.92/3.12 5.12/3.24
CoDE EPSDE SFMDE CDE
F1 450.000 7.43E8 2/11/0 450.000 9.46E8 2/11/0 449.974 1.43E1 1/0/12 450.000 1.76E7 2/11/0
F2 450.000 2.34E6 3/10/0 450.000 2.00E8 3/10/0 355.757 5.16E+2 1/0/12 450.000 6.14E8 3/10/0
F3 2.32E+05 1.12E+5 6/3/4 3.53E+05 1.82E+5 4/2/7 3.75E+06 9.85E+6 1/4/8 1.37E+06 5.36E+5 2/0/11
F4 449.418 6.45E1 5/1/7 449.998 3.79E3 8/2/3 1654.396 3.17E+3 1/0/12 450.000 1.56E5 11/2/0
F5 87.820 2.04E+2 8/1/4 258.714 6.01E+2 8/1/4 2865.587 1.09E+3 1/0/12 1520.590 4.55E+2 3/1/9
F6 403.509 1.12E+1 3/0/10 390.133 7.28E1 11/2/0 520.362 1.71E+2 1/1/11 397.526 1.16E+0 6/0/7
F7 179.972 5.57E2 3/4/6 179.997 5.09E3 12/0/1 175.777 1.31E+1 1/0/12 179.995 5.89E3 8/3/2
F8 119.049 4.58E2 0/11/2 119.048 5.24E2 0/11/2 119.293 1.64E1 12/0/1 119.060 4.78E2 0/11/2
F9 330.000 7.32E8 10/3/0 304.132 2.33E+0 3/1/9 306.477 1.02E+1 3/1/9 330.000 1.90E7 10/3/0
F10 154.278 1.30E+1 1/0/12 168.704 9.19E+0 2/0/11 244.334 4.94E+1 5/0/8 225.003 1.11E+1 4/0/9
F11 122.765 1.53E+0 3/1/9 125.424 1.43E+0 1/0/12 122.566 2.23E+0 3/1/9 121.327 1.40E+0 5/1/7
F12 8.71E+04 1.33E+4 1/1/11 6.49E+04 4.60E+4 1/1/11 2.42E+04 3.96E+4 4/2/7 5.75E+04 9.88E+3 3/0/10
F13 123.940 6.99E1 3/0/10 123.478 6.99E1 1/0/12 127.080 7.69E1 8/2/3 126.509 2.93E1 6/1/6
F14 286.790 1.95E1 1/3/9 286.743 2.10E1 1/2/10 286.844 2.62E1 1/4/8 286.856 1.85E1 2/3/8
F15 526.667 5.21E+1 0/8/5 510.000 1.09E+2 1/11/1 525.053 4.02E+1 1/1/11 503.333 6.48E+1 1/10/2
F16 312.983 2.06E+1 1/1/11 353.324 9.16E+1 1/1/11 254.364 5.58E+1 3/3/7 238.450 1.22E+1 4/3/6
F17 357.225 2.79E+1 1/1/11 399.793 9.39E+1 0/2/11 265.719 5.02E+1 4/4/5 273.235 2.68E+1 5/2/6
F18 916.093 2.41E1 2/4/7 914.921 1.28E+0 4/4/5 911.417 3.55E+1 6/6/1 915.487 6.10E1 3/4/6
F19 915.975 3.43E1 2/3/8 914.200 8.78E1 6/4/3 916.077 3.00E+1 1/2/10 915.328 6.18E1 3/2/8
F20 915.893 3.08E1 1/4/8 914.694 1.56E+0 4/6/3 910.145 2.74E+1 8/5/0 915.551 9.70E1 2/3/8
F21 860.000 1.1E13 4/9/0 870.000 5.48E+1 0/4/9 860.022 1.2E1 2/0/11 860.000 2.2E13 4/9/0
F22 1262.959 9.76E+0 2/2/9 1242.860 1.54E+1 6/6/1 1325.995 2.93E+1 0/2/12 1256.220 1.38E+1 2/3/8
F23 894.165 2.72E4 6/7/0 907.569 7.34E+1 2/1/10 894.251 4.72E1 2/2/9 894.164 7.53E5 5/8/0
F24 460.000 1.2E12 1/12/0 460.000 0.00E+0 1/12/0 460.000 4.2E10 1/12/0 460.000 0.00E+0 1/12/0
F25 470.800 8.25E1 7/2/3 470.416 7.01E1 8/4/1 475.174 4.47E+0 2/0/11 471.486 6.52E1 3/4/6
H/APv 4.21E7 (++) 5.13E5 (++) 3.65E8 (++) 0.0055 (++)
W/Pav 3.04/5.88 3.60/5.48 2.92/8.04 3.92/4.84
1
8
0
A
.
P
.
P
i
o
t
r
o
w
s
k
i
/
I
n
f
o
r
m
a
t
i
o
n
S
c
i
e
n
c
e
s
2
4
1
(
2
0
1
3
)
1
6
4
1
9
4
Table 4
The mean tness values (mean) and its standard deviations (std) for 50-dimensional problems. +/=/ shows than how many of 13 competing methods the particular algorithm is signicantly better/equal/
signicantly worse, according to Wilcoxon rank-sum test at 5% signicance level. W/Pav how many times, averaged over all problems, the particular algorithm is signicantly better/signicantly worse than other
algorithms according to Wilcoxon test. H/APv adjusted p-values according to Holm test procedure with AM-DEGL as a control method; ++ means that AM-DEGL is signicantly better than the competitive algorithm,
== means that there is no signicant difference between AM-DEGL and competitive method at 5% signicance level, according to Holms procedure.
AM-DEGL DE DE-SG DEGL SADE
Mean Std +/=/ Mean Std +/=/ Mean Std +/=/ Mean Std +/=/ Mean Std +/=/
F1 450.000 2.64E7 2/11/0 293.680 7.95E+1 0/0/13 449.902 4.20E2 1/0/12 450.000 8.94E8 2/11/0 450.000 5.43E8 2/11/0
F2 450.000 1.29E7 11/2/0 8.58E+04 6.94E+3 0/0/13 428.595 1.44E+1 3/1/9 450.000 2.47E8 11/2/0 450.000 1.39E5 11/2/0
F3 3.89E+04 5.35E+4 13/0/0 5.37E+08 6.83E+7 0/0/13 1.03E+06 4.40E+5 5/2/6 1.86E+05 8.38E+4 11/1/1 1.84E+05 7.00E+4 11/1/1
F4 418.021 1.05E+2 12/0/1 1.05E+05 9.37E+3 0/0/13 319.150 3.72E+2 5/2/6 449.998 3.67E3 13/0/0 237.338 2.76E+2 10/1/2
F5 305.269 5.32E+2 13/0/0 1.78E+04 8.63E+2 0/0/13 3463.695 6.42E+2 5/3/5 1231.147 6.45E+2 12/0/1 2120.398 5.70E+2 11/0/2
F6 399.550 9.89E+0 12/0/1 3.59E+07 5.40E+6 0/0/13 897.727 1.08E+2 1/0/12 390.930 1.71E+0 13/0/0 436.096 3.60E+1 6/4/3
F7 179.995 7.93E3 7/4/2 3999.297 2.59E+2 0/0/13 173.131 4.40E+0 2/0/11 179.994 9.41E3 6/7/0 179.994 1.15E2 5/8/0
F8 119.945 3.43E2 13/0/0 118.866 4.51E2 0/11/2 118.872 3.61E2 0/11/2 118.873 2.14E2 0/11/2 118.868 5.44E2 0/11/2
F9 296.287 7.61E+0 7/0/6 21.667 1.10E+1 0/0/13 265.928 5.89E+0 3/2/8 265.892 1.57E+1 3/2/8 329.998 2.47E3 12/0/1
F10 256.589 3.88E+1 13/0/0 125.001 1.49E+1 0/0/13 57.282 1.72E+1 4/1/8 126.071 1.28E+2 4/5/4 188.654 3.34E+1 6/1/6
F11 134.273 5.97E+0 12/1/0 162.863 1.39E+0 0/0/13 150.797 1.36E+0 5/1/7 131.208 2.42E+1 7/6/0 146.633 1.49E+0 9/2/2
F12 1.08E+04 8.73E+3 9/4/0 2.01E+06 1.38E+5 0/0/13 9.56E+04 4.94E+4 3/1/9 2.41E+04 2.28E+4 4/4/5 1.21E+04 8.90E+3 7/6/0
F13 125.660 9.39E1 12/0/1 69.971 3.99E+0 0/0/13 120.468 5.64E1 6/1/6 109.071 8.59E+0 1/0/12 122.097 4.04E1 7/2/4
F14 277.630 4.05E1 8/3/2 276.851 1.43E1 0/1/12 277.371 2.24E1 6/2/5 277.455 3.06E1 6/4/3 277.456 1.60E1 7/3/3
F15 440.206 9.63E+1 6/6/1 620.380 6.24E+0 0/0/13 520.016 8.31E3 1/0/12 441.202 8.49E+1 5/7/1 506.701 5.08E+1 2/2/9
F16 181.144 2.29E+1 12/0/1 450.983 8.81E+0 0/0/13 300.083 6.28E+0 4/1/8 297.440 1.34E+2 4/7/2 218.645 2.18E+1 7/4/2
F17 276.789 8.08E+1 11/2/0 486.354 1.21E+1 0/0/13 363.241 1.34E+1 5/0/8 388.075 8.66E+1 2/2/9 293.431 1.56E+1 7/2/4
F18 927.578 2.75E+0 7/0/6 967.641 2.60E+0 0/0/13 922.038 3.17E+1 7/3/3 913.364 1.03E+2 9/3/1 916.752 3.87E+1 7/5/1
F19 927.248 2.63E+0 8/0/5 968.580 2.64E+0 0/0/13 918.738 3.68E+1 10/3/0 928.048 2.54E+1 5/3/5 920.103 3.75E+1 9/4/0
F20 923.279 2.16E+1 11/2/0 967.125 2.94E+0 0/0/13 929.250 2.24E+1 4/4/5 931.005 1.40E+1 3/4/6 923.335 3.12E+1 5/6/2
F21 860.000 2.7E13 8/5/0 1392.258 8.55E1 0/0/13 860.009 3.96E3 5/2/6 948.999 1.87E+2 4/1/8 860.000 3.5E13 8/5/0
F22 1265.103 1.11E+1 11/2/0 1319.339 4.74E+0 0/1/12 1298.403 1.46E+1 2/2/9 1304.029 2.37E+1 0/5/8 1277.948 1.84E+1 7/3/3
F23 899.122 1.62E4 9/4/0 1384.574 9.87E1 0/0/13 899.126 4.39E3 8/1/4 990.950 1.62E+2 5/0/8 899.122 6.07E5 9/4/0
F24 460.000 0.00E+0 3/10/0 727.428 3.46E+1 1/0/12 481.296 1.08E+2 2/0/11 460.000 1.6E12 3/10/0 460.000 5.9E14 3/10/0
F25 474.191 8.22E1 13/0/0 924.502 3.51E+1 0/0/13 484.122 2.85E+0 3/0/10 477.374 1.70E+0 9/1/3 475.107 9.41E1 11/1/1
H/APv - 2.67E-21 (++) 7.23E-7 (++) 0.0047 (++) 0.0327 (++)
W/Pav 9.72/1.04 0.04/12.44 4.00/7.28 5.68/3.48 7.16/1.92
SspDE JADE AdapSS-JADE DECLS DEahcSPX
F1 450.000 3.05E8 2/11/0 450.000 6.28E8 2/11/0 450.000 6.72E8 2/11/0 450.000 4.18E8 2/11/0 450.000 3.86E8 2/11/0
F2 449.462 3.22E1 6/0/7 449.942 1.34E1 8/2/3 448.748 1.24E+0 5/0/8 449.762 4.89E1 7/0/6 164.376 2.83E+2 2/0/11
F3 6.45E+05 2.92E+5 8/1/4 1.15E+06 3.66E+5 4/3/6 1.14E+06 2.99E+5 4/3/7 7.31E+05 2.81E+5 8/1/4 4.56E+06 1.42E+6 2/0/11
F4 1361.086 9.65E+2 4/0/9 101.276 3.53E+2 8/0/5 374.127 4.33E+2 5/2/6 556.923 7.45E+2 5/2/6 4786.467 1.77E+3 2/1/10
F5 5361.632 8.37E+2 3/0/10 3632.487 5.43E+2 5/2/6 2985.857 3.70E+2 8/2/3 3641.678 5.92E+2 5/2/6 2747.168 5.82E+2 9/1/3
F6 457.461 3.19E+1 4/1/8 425.684 2.16E+1 6/3/4 426.804 2.16E+1 7/2/4 427.663 3.02E+1 6/2/5 509.278 1.15E+2 3/2/8
F7 179.995 9.33E3 5/6/2 179.998 6.08E3 8/5/0 179.998 5.95E3 8/5/4 179.991 1.37E2 5/4/4 179.984 1.68E2 4/1/8
F8 118.869 3.78E2 0/11/2 118.870 3.31E2 0/11/2 118.867 3.37E2 0/11/2 118.869 3.53E2 0/11/2 118.862 3.15E2 0/11/2
F9 330.000 3.99E8 12/1/0 174.367 9.92E+0 1/0/12 326.518 1.75E+0 10/0/3 329.702 6.99E1 11/1/1 288.378 1.43E+1 6/0/7
F10 254.397 6.61E+0 10/2/1 196.578 1.32E+1 7/2/4 195.907 1.32E+1 7/2/4 246.409 1.83E+1 10/2/1 17.705 3.01E+1 3/0/10
F11 143.870 1.84E+0 11/1/1 148.883 1.70E+0 7/2/4 149.070 1.44E+0 7/2/4 146.441 2.29E+0 9/2/2 161.354 1.87E+0 1/0/12
F12 2.25E+04 1.34E+4 4/4/5 9394.746 1.06E+4 9/4/0 1.15E+04 1.07E+4 9/4/0 1.28E+04 1.31E+4 7/6/0 1.98E+04 1.72E+4 5/5/3
F13 125.415 2.94E1 10/1/2 115.983 8.89E1 4/1/8 116.256 1.12E+0 4/1/8 127.391 2.01E1 13/0/0 119.892 5.65E+0 6/3/4
F14 277.404 2.12E1 6/3/4 278.021 2.38E1 12/1/0 277.908 1.99E1 12/1/0 277.683 3.11E1 10/1/2 277.031 2.44E1 1/3/9
(continued on next page)
A
.
P
.
P
i
o
t
r
o
w
s
k
i
/
I
n
f
o
r
m
a
t
i
o
n
S
c
i
e
n
c
e
s
2
4
1
(
2
0
1
3
)
1
6
4
1
9
4
1
8
1
Table 4 (continued)
AM-DEGL DE DE-SG DEGL SADE
Mean Std +/=/ Mean Std +/=/ Mean Std +/=/ Mean Std +/=/ Mean Std +/=/
F15 490.695 7.89E+1 3/3/7 446.667 9.80E+1 6/6/1 463.333 8.98E+1 5/6/2 407.348 9.28E+1 10/3/0 454.450 9.48E+1 5/3/5
F16 172.594 2.02E+1 13/0/0 219.504 1.78E+1 7/4/2 220.772 7.47E+0 6/4/3 219.763 6.52E+1 7/2/4 344.450 2.91E+1 3/0/10
F17 255.686 2.40E+1 11/2/0 283.548 1.10E+1 8/3/2 281.919 2.05E+1 8/3/2 295.089 3.33E+1 7/4/2 379.880 2.11E+1 3/1/9
F18 918.215 5.52E+1 10/0/3 930.791 8.44E+0 5/3/5 929.397 1.32E+0 5/1/7 932.074 3.49E+0 3/4/6 925.361 1.70E+0 8/1/4
F19 924.607 4.61E+1 11/0/2 929.138 3.81E+0 5/2/6 929.243 2.25E+0 5/2/6 932.998 3.07E+0 3/2/8 925.328 1.32E+0 9/1/3
F20 927.503 4.11E+1 9/0/4 928.498 4.39E+0 7/2/4 929.986 2.12E+0 6/2/5 932.430 2.73E+0 3/4/6 925.810 1.67E+0 10/2/1
F21 860.000 5.8E13 6/7/0 860.000 3.5E13 8/5/0 860.000 3.5E13 8/5/0 1113.801 2.58E+2 2/5/6 1280.999 1.92E+2 2/1/10
F22 1283.716 2.05E+1 5/5/3 1271.691 1.55E+1 10/3/0 1272.152 1.28E+1 7/5/1 1268.522 1.42E+1 10/3/0 1294.279 8.88E+0 2/2/9
F23 899.133 1.57E2 8/4/1 914.802 8.59E+1 7/1/5 899.122 3.43E5 9/4/0 1245.744 2.13E+2 3/1/9 1342.837 1.12E+2 2/0/11
F24 460.000 1.1E12 3/10/0 460.000 9.7E13 3/10/0 460.000 5.8E14 3/10/0 460.000 1.4E12 3/10/0 460.000 5.10E9 3/10/0
F25 479.342 2.17E+0 4/4/5 479.091 1.11E+0 4/4/5 478.766 6.27E1 4/4/5 478.081 2.07E+0 9/1/3 479.416 1.92E+0 4/4/5
H/APv 0.0052 (++) 0.0055 (++) 0.0055 (++) 0.0025 (++) 2.73E8 (++)
W/Pav 6.72/3.20 6.04/3.36 6.16/3.16 6.32/3.32 3.88/6.60
CoDE EPSDE SFMDE CDE
F1 450.000 4.89E8 2/11/0 450.000 6.93E8 2/11/0 450.000 3.80E6 2/11/0 450.000 1.36E7 2/11/0
F2 375.346 4.22E+1 3/0/10 449.974 2.00E2 9/1/3 870.534 7.13E+3 1/1/11 449.971 8.67E3 8/1/4
F3 2.86E+06 1.09E+6 3/0/10 1.38E+06 6.46E+5 4/2/7 4.68E+05 4.80E+5 10/0/3 2.79E+07 7.14E+6 1/0/12
F4 4932.250 1.89E+3 2/1/10 114.050 2.53E+2 9/0/4 1.59E+04 1.00E+4 1/0/12 367.831 3.02E+1 10/1/2
F5 1.12E+04 9.84E+2 1/0/12 3288.718 7.86E+2 7/2/4 6821.534 1.62E+3 2/0/11 4774.867 6.44E+2 4/0/9
F6 702.219 4.95E+1 3/0/10 409.735 1.79E+1 11/0/2 729.542 4.40E+2 2/1/10 423.880 5.55E1 9/1/3
F7 177.018 7.26E1 3/0/10 179.996 6.85E3 5/8/0 169.119 2.78E+1 1/8/4 179.932 1.32E2 4/0/9
F8 118.879 4.10E2 0/11/2 118.868 3.55E2 0/11/2 119.208 2.35E1 12/0/1 118.870 3.34E2 0/11/2
F9 299.747 2.19E+0 8/0/5 203.869 7.49E+0 2/0/11 254.396 4.73E+1 3/2/8 306.352 1.65E+0 9/0/4
F10 88.301 1.64E+1 1/0/12 9.460 1.19E+1 2/0/11 225.316 7.23E+1 10/2/1 111.091 1.73E+1 5/1/7
F11 157.221 1.46E+0 3/0/10 159.728 1.43E+0 2/0/11 155.391 2.80E+0 4/0/9 150.828 2.10E+0 5/1/7
F12 6.31E+05 6.41E+4 1/0/12 2.03E+05 2.31E+5 2/5/6 6.48E+04 2.27E+5 4/6/3 3.01E+05 3.14E+4 2/1/10
F13 109.704 1.63E+0 2/0/11 112.030 1.00E+0 3/0/10 125.188 1.01E+0 10/1/2 122.019 4.03E1 7/2/4
F14 276.938 1.42E1 1/3/9 276.924 1.74E1 0/4/9 276.985 2.36E1 1/3/9 277.232 1.89E1 5/0/8
F15 506.673 5.07E+1 3/1/9 450.000 9.52E+1 6/7/0 433.764 8.59E+1 3/8/2 423.896 1.00E+2 7/6/0
F16 414.369 1.71E+1 1/0/12 394.278 5.99E+1 2/0/11 236.067 7.13E+1 6/5/2 271.471 1.30E+1 5/1/7
F17 452.550 1.32E+1 1/0/12 405.291 4.62E+1 2/1/10 284.295 8.74E+1 7/6/0 325.528 1.05E+1 6/0/7
F18 931.994 9.00E1 3/3/7 876.721 1.57E+2 12/1/0 949.578 3.00E+1 1/0/12 936.087 1.98E+0 2/0/11
F19 931.634 9.67E1 4/1/8 925.393 3.46E+0 9/1/3 945.795 3.00E+1 1/0/12 935.784 2.22E+0 2/1/10
F20 931.579 8.76E1 3/3/7 885.407 1.56E+2 11/2/0 951.914 1.22E+1 1/0/12 936.414 1.85E+0 2/1/10
F21 1346.459 1.32E+2 1/0/12 1145.780 2.54E+2 3/5/5 924.787 1.12E+2 4/2/7 860.000 3.5E13 8/5/0
F22 1303.479 5.82E+0 1/1/11 1280.048 1.45E+1 6/2/5 1283.507 2.09E+1 6/4/3 1289.949 7.02E+0 4/2/7
F23 1377.713 9.03E1 1/0/12 1276.835 1.92E+2 3/1/9 929.046 7.76E+1 6/0/7 899.123 1.11E4 9/4/0
F24 460.000 1.06E6 3/10/0 460.000 5.8E14 3/10/0 768.498 1.20E+2 0/0/13 460.000 5.8E14 3/10/0
F25 479.724 2.33E+0 4/4/5 475.036 1.06E+0 11/1/1 733.659 1.61E+2 1/0/12 514.958 4.05E+0 2/0/11
H/APv 4.60E12 (++) 4.79E5 (++) 3.04E8 (++) 6.65E5 (++)
W/Pav 2.32/8.72 5.04/4.96 3.96/6.64 4.84/5.76
1
8
2
A
.
P
.
P
i
o
t
r
o
w
s
k
i
/
I
n
f
o
r
m
a
t
i
o
n
S
c
i
e
n
c
e
s
2
4
1
(
2
0
1
3
)
1
6
4
1
9
4
Table 5
The mean tness values of different variants of the proposed algorithm on 30-dimensional problems. W/Pav informs how many times, averaged over all problems, the particular algorithm is signicantly better/
signicantly worse than other algorithms according to Wilcoxon test.
AM-
DEGL
F-SADE F-xed RA-NMA RA SADE-
glob-only
LM-
greedy
Loc-Only-
best
Loc-
every
p-strat-
from-g2
p-strat-
SADE
p
6
p 10 p 40 NMA
-closest
Mean
F1 450.00 450.00 450.00 450.00 450.00 450.00 450.00 450.00 450.00 450.00 450.00 450.00 450.00 450.00 450.00
F2 450.00 450.00 450.00 450.00 450.00 450.00 450.00 450.00 450.00 450.00 450.00 450.00 450.00 450.00 450.00
F3 1780.31 7948.30 312.35 308.01 2382.32 120.69 10130.32 6.16 301.46 721.97 350.44 202.19 12.44 1319.68 29.59
F4 450.00 285.52 450.00 450.00 450.00 450.00 450.00 450.00 450.00 450.00 450.00 449.87 449.97 450.00 450.00
F5 309.98 1215.05 310.00 308.33 307.92 309.64 309.80 309.98 309.99 309.97 309.93 309.28 309.77 310.00 309.99
F6 391.59 391.77 392.70 399.28 397.60 401.90 390.80 392.67 392.26 390.63 392.01 394.48 392.68 392.98 391.04
F7 180.00 179.97 180.00 179.99 179.99 179.99 179.99 179.99 180.00 180.00 180.00 179.99 180.00 180.00 179.99
F8 119.98 119.98 119.98 119.87 119.67 119.99 119.98 120.00 119.98 119.98 119.98 120.00 119.99 119.96 119.99
F9 326.66 328.03 324.30 321.47 324.00 329.48 329.91 326.90 326.00 326.30 326.29 320.43 321.56 329.49 327.98
F10 297.07 274.12 295.64 285.94 271.34 291.99 296.07 272.96 297.89 294.50 299.39 294.68 293.92 290.26 287.91
F11 113.84 112.02 113.21 115.93 118.85 111.69 107.29 113.22 111.87 112.76 112.31 112.47 112.60 113.77 115.97
F12 1915.83 2274.20 2815.13 5539.57 8293.43 2699.04 4713.38 3901.60 2020.65 2515.67 3248.81 3052.01 2279.87 1904.84 1511.68
F13 128.03 127.99 128.03 127.87 127.59 128.09 128.03 127.98 128.20 128.12 127.99 128.02 128.07 127.94 128.19
F14 287.29 287.88 287.35 287.21 287.31 287.22 287.56 287.27 287.37 287.22 287.27 287.19 287.29 287.54 287.34
F15 520.00 524.18 520.00 500.07 510.08 523.52 470.00 506.67 506.72 540.07 516.67 490.07 507.69 513.49 490.07
F16 197.90 230.07 211.29 217.58 205.05 193.94 225.04 242.83 229.82 225.04 208.91 229.18 210.15 246.72 212.56
F17 220.79 241.13 239.01 258.37 340.39 219.28 222.95 269.48 235.35 216.26 231.61 226.02 228.86 221.74 272.38
F18 914.44 887.72 914.65 911.33 914.95 911.52 914.58 904.40 897.27 911.21 911.18 908.13 907.80 911.12 911.23
F19 900.70 887.75 907.70 911.48 914.70 908.06 914.33 911.30 910.83 907.43 907.72 904.87 900.51 911.48 908.03
F20 914.88 914.11 914.70 914.88 914.64 911.56 914.70 911.36 914.53 907.81 911.18 901.08 908.07 914.67 908.35
F21 860.00 911.91 870.00 860.00 860.00 860.00 860.00 860.00 860.00 880.00 860.00 860.00 880.00 870.00 860.00
F22 1230.94 1262.75 1225.23 1234.46 1233.34 1237.18 1235.19 1231.76 1235.04 1235.36 1228.29 1233.73 1235.62 1229.85 1229.81
F23 894.16 894.18 894.16 894.16 894.16 894.16 907.59 894.16 894.16 894.16 894.16 921.02 894.16 894.16 894.16
F24 460.00 460.00 460.00 460.00 460.00 460.00 460.00 460.00 460.00 460.00 460.00 460.00 460.00 460.00 460.00
F25 469.66 474.93 469.39 469.92 469.65 469.53 471.02 469.65 469.67 469.78 469.65 470.18 469.92 469.51 469.82
W/Pav 2.68/1.04 3.20/5.64 2.88/1.36 1.28/4.56 1.24/5.72 2.68/2.04 3.44/2.52 2.16/1.96 2.60/1.32 3.08/1.20 2.40/1.12 2.32/ 3.00 2.48/2.08 2.68/2.08 2.64/2.12
A
.
P
.
P
i
o
t
r
o
w
s
k
i
/
I
n
f
o
r
m
a
t
i
o
n
S
c
i
e
n
c
e
s
2
4
1
(
2
0
1
3
)
1
6
4
1
9
4
1
8
3
The +/=/ values refer to: + the number of algorithms that perform signicantly worse than the considered algorithm;
= the number of algorithms with performance that does not differ signicantly from the performance of the considered
algorithm; the number of algorithms that perform signicantly better than the considered algorithm. Note that uni-
modal problems F1 and F2 are easily solved by most DE approaches. It was found that AM-DEGL is among the quickest meth-
od for such problems, hence it could be awarded by additional + just for quick exploitation. Hence, to not overestimate the
performance of the proposed method, it was ultimately decided that in this paper when Wilcoxon rank-sum test is used the
performance of two algorithms is considered equal (=) if their mean results differ by less than 10
3
.
F3
0E00
3.E+06
6.E+06
9.E+06
fitn
e
s
s
AM-DEGL DE DE-SG DEGL
SADE SspDE JADE AdapSS-JADE
DECLS DEahcSPX CoDE EPSDE
SFMDE CDE
F3
0.E+00
3.E+06
6.E+06
9.E+06
0 100000 200000 300000
n.f.c.
f
i
t
n
e
s
s
F10
-350
-200
-50
100
0 100000 200000 300000
n.f.c.
f
i
t
n
e
s
s
F12
0.E+00
1.E+05
2.E+05
3.E+05
0 100000 200000 300000
n.f.c.
f
i
t
n
e
s
s
F17
200
400
600
0 100000 200000 300000
n.f.c.
f
i
t
n
e
s
s
F19
880
930
980
n.f.c.
f
i
t
n
e
s
s
F22
1200
1300
1400
1500
0 100000 200000 300000
0 100000 200000 300000
n.f.c.
f
i
t
n
e
s
s
Fig. 3. Mean convergence characteristics of 14 compared algorithms for selected 30-dimensional problems.
184 A.P. Piotrowski / Information Sciences 241 (2013) 164194
In Tables 24 for each problem only the results obtained by the algorithm that is signicantly better according to Wilcoxon
test than all 13 competing methods are bolded, what is rare. Frequently at least two out of a large number of algorithms
perform similarly for a given problem and the difference in their performance is not statistically signicant the difculty
of nding an algorithm that perform better than all others for many problems is also seen in comparisons presented in
[101,94,43,36,44,54]. With respect to that in the last row in Tables 24 (denoted W/Pav) for every algorithm the averaged
over all problems number of DE variants that perform, according to Wilcoxon rank-sum test, worse (+) / better () then
the considered algorithm is shown.
According to rank-based Holm procedure in case of 10-dimensional problems AM-DEGL is signicantly better (at 5% sig-
nicance level) than 5 competitive methods. The proposed AM-DEGL is neither signicantly better nor signicantly worse
than other 8 algorithms. When 30-dimensional problems are compared, AM-DEGL is signicantly better than 12 methods,
SADE is an exception. However, at 10% signicance level (not shown in Table 3) AM-DEGL is signicantly better than all 13
other DE variants. In case of 50-dimensional problems AM-DEGL is signicantly better than all competitive algorithms at 5%
signicance level.
However, when Wilcoxon test is applied for inter-comparison between all methods for each problem separately, usually
the differences between best algorithms are not statistically signicant. Only for 3 out of 25 10-dimensional problems some
algorithms (AM-DEGL, DE-SG and DEGL, each only once) signicantly outperform all other methods. In case of 30-dimen-
sional problems AM-DEGL and DEGL signicantly outperform all other algorithms for 3 and 2 problems, respectively. For
50-dimensional problems AM-DEGL signicantly outperforms all other algorithms for 5 problems, DEGL and SspDE for one.
Comparing the averaged over all problems number of algorithms that perform worse/better than the considered algo-
rithm by means of Wilcoxon test, the best results for 10-dimensional problems are obtained by means of DE-SG (it is on aver-
age signicantly better than 6.76 algorithms) and SADE (it is on average signicantly worse than 0.84 algorithms). AM-DEGL
nish the second winning (6.24) and second the most rarely loosing (1.6) method. For 30-dimensional problems AM-DEGL is
the best method, on average better than 8.00 and worse than 1.24 algorithms. The second-best method (SADE) is on average
better than 6.12 and worse than 1.84 algorithms. For 50-dimensional problems AM-DEGL is also the best algorithm, on aver-
age better than 9.72 and worse than 1.04 competitors. The second best method (SADE) is on average better than 7.16 and
worse than 1.92 algorithms.
One may note that the higher dimensionality, the larger difference between AM-DEGL and other DE algorithms, both
according to rank-based Holm procedure and Wilcoxon test. Considering 30- and 50-dimensional variants of CEC 2005
problems, AM-DEGL performance is among the best methods for all test functions except F9, F18 and F19. For none of
considered problems AM-DEGL is signicantly worse than 6 out of 13 algorithms. Among 30- and 50-dimensional prob-
lems AM-DEGL performance is not signicantly worse than any among 13 algorithms for 14 and 15 test functions,
respectively.
It is not easy to answer the question for what kind of problems AM-DEGL performance is the best, or for which problems
is the poorest. For uni-modal problems (F1F5) AM-DEGL and DEGL are the two best methods, irrespective of problem
dimensionality. This indicates that AM-DEGL inherited the possibility of quick detection of the minimum from DEGL, one
of its ancestor methods, enhanced by using a local search algorithm. In case of basic multimodal functions AM-DEGL perfor-
mance is diversied for 50-dimensional problems F8 and F10 and 30-dimensional problems F7 and F8 it is signicantly
better than all other methods. Its performance is also among the best methods for F11 and F12, irrespective of dimension-
ality. However, comparing to other DE variants, AM-DEGL does not work well for 10-dimensional variant of F6 and especially
for problemF9 (shifted Rastrigins function), irrespective of dimensionality. One may note that for F9 problemboth DEGL and
SFMDE (which, like AM-DEGL, also uses NMA algorithm as local search method) also perform poor, while performance of
SADE is among the best. This indicate that for non-rotated shifted Rastrigins function the more exploitative features of
AM-DEGL may lead to the premature convergence. However, when rotated variant of shifted Rastrigins function is used
(problem F10), the performance of AM-DEGL is the best among tested methods, especially for higher-dimensional problems,
conrming good balance between time-consuming adaptive approaches that strength exploration, and exploitation capabil-
ities inherited from DEGL and SFMDE. For expanded multimodal (F13F14) and composition functions (F15F25) the perfor-
mance of AM-DEGL is among the best methods, with exceptions of 10- and 30-dimensional variants of problems F14, F18,
F19 and F20, and 50-dimensional variants of problems F18 and F19, for which AM-DEGL ranked middle positions among DE
algorithms. However, note that composition problems F18, F19 and F20 are composed from the same functions with differ-
ent parameterization, and for this three problems none of parent methods of AM-DEGL perform well. On the other hand,
for seven among ten 50-dimensional composition functions no algorithm performs better than AM-DEGL. It also seems that
AM-DEGL is not as signicantly affected by rotation as other DE variants according to Wilcoxon test, AM-DEGL outperform
other DE variants much more frequently for shifted rotated Rastrigins function (F10) than for shifted Rastrigins function
(F9) and for rotated composition function (F16) than its non-rotated variant (F15), especially when 30- and 50-dimensional
versions are considered. Such behavior is not observed for DEGL or SADE. As most DE methods are known for its weak ability
to cope with rotated problems due to the nature of crossover, the better performance for rotated problems by means of AM-
DEGL may be the reason of two factors the use of NMA as local search method (note that the performance of SFMDE, com-
paring to other DE variants, is also relatively better for problems F10 and F16, then for F9 and F15), or the quick adaptation of
probability of using different mutation strategies (of which one is not followed by crossover; although SADE also uses this
strategy, the adaptation of probability of using different strategies in SADE is much slower than in AM-DEGL).
A.P. Piotrowski / Information Sciences 241 (2013) 164194 185
Fig. 3 shows the convergence characteristics of all compared algorithms for 6 selected 30-dimensional test functions. It
may be seen that AM-DEGL is among the quickest methods for each problem. It is frequently slightly slower only to DEGL,
but in vast majority of cases it reaches signicantly better results than all its ancestral approaches.
The obtained results, especially for 30- and 50-dimensional problems, may suggest that AM-DEGL is more reliable meth-
od than its ancestral algorithms and other DE variants (but note the discussion in Section 6). In the future it is planned to test
AM-DEGL on high-dimensional problems. However, this may require a bit different settings of the population size, as in DE
algorithms used for large scale optimization usually NP D [145,157,16]. As a result some novel method will be needed to
determine D + 1 points to be used by NMA simplex, what may lead to an important modication of AM-DEGL.
5.4. Motivation of different parts of AM-DEGL
In this sub-section it is veried howthe performance of AM-DEGL depends on some of its components and control param-
eter values. The comparison of AM-DEGL with a number of its modied variants is presented for 30-dimensional CEC 2005
problems. The following modications of AM-DEGL algorithm dened in Section 4 are considered:
1. F-SADE the scale factor F
i,g
is generated from N(0.5, 0.3), following SADE;
2. F-xed the scale factor F
i,g
is xed to 0.8, following DEGL;
3. RA-NMA two local search methods, RA [117] and NMA may be used by each individual that is better than all its
neighbors, each with the same probability (LSp = 0.005). The maximum number of function calls by each local search
method is kept 100D as in AM-DEGL. In this variant the probability of commencing local search and the average num-
ber of function calls used by the local search methods is twice higher than in AM-DEGL;
4. RA only RA is used as local search method, instead of NMA. The probability of commencing local search and the max-
imum number of function calls used by local search algorithm are the same as in AM-DEGL;
5. SADE-glob-only the Local mutation model is not used by four strategies that were applied in SADE, it is used by
DEGL-based strategy only;
6. LM-greedy irrespective of Global mutation strategy, as Local mutation model the most greedy DE/best/1 strategy
(Eq. (4)) is applied, where instead of x
best,g
the x
Lbest,i,g
is used. This variant is considered to check the possible advan-
tage of highly exploitative local mutation model;
7. Loc-only-best the local search may be commenced by the best individual in the population only, following [98].
Because on average 45 individuals have a chance to commence local search by AM-DEGL for 30-dimensional prob-
lems, the probability of commencing local search in this variant is increased by a factor of 4 (LSp = 0.02), to keep the
number of local searches per run roughly similar to the original AM-DEGL;
8. Loc-every the local search may be performed by each individual in the population. Because on average 45 individ-
uals have a chance to commence local search by AM-DEGL for 30-dimensional problems, the probability of commenc-
ing local search by this variant is divided by 30 (LSp = 0.000167), to keep the number of local searches per run roughly
similar to the original AM-DEGL;
9. p-start-from-g2 probabilities of using different strategies are adapted from the second generation, without warm-
up period;
10. p-start-SADE probabilities of using different strategies are adapted following equations proposed in SADE (Eq. (14) in
[115] with LP = 1), instead of Eqs. (22) and (23);
11. p 6 the value of p% neighborhood is reduced to 6%;
12. p 10 the value of p% neighborhood proposed in DEGL is used (i.e. 10%);
13. p 40 the value of p% neighborhood is increased to 40%;
14. NMA-closest a different procedure of choosing the initial D + 1 points for NMA algorithm is used. In this variant the
x
Lbest,i,g
and its D closest individuals in the search space (according to Euclidean norm) are chosen. The motivation
beyond this scenario is that such chosen D + 1 points may facilitate quick exploitation of the local optimum. However,
the explorative features of NMA are not well exploited in this procedure.
The mean performance of AM-DEGL and its 14 variants are presented in Table 5. For each version the averaged over all
problems number of other variants that are, according to Wilcoxon test, statistically signicantly better/worse is presented
in W/Pav row. It is seen that on average the performances of AM-DEGL dened in Section 4 and most of its variants are com-
parable. As in each considered variant only one component of AM-DEGL has been modied, this conrms that the small
changes to AM-DEGL do not destroy its ability to nd the optima of various problems. It also conrms that the good perfor-
mance of AM-DEGL is not an effect of a single novelty proposed, but rather a successful combination of well known elements
of already developed DE variants, what was the intention of the present paper. However, a few AM-DEGL variants perform
poorer than the others:
1. The use of greedy local search RA algorithm instead of, or together with NMA (RA and RA-NMA) disturbs the search ability
of AM-DEGL. Although benecial in some cases (F3, F15, F18) it usually diminish the exploration capabilities of the
method.
186 A.P. Piotrowski / Information Sciences 241 (2013) 164194
2. The method of generating scale factor borrowed from SADE (F-SADE) turns out un-trustworthy when applied in AM-
DEGL. Although the number of variants from which F-SADE is signicantly better than other AM-DEGL variants is the
largest (also larger than AM-DEGL), the number of variants that are signicantly better than F-SADE is also much higher
than noted for any other variant. This suggests that the scale factor generation method used in SADE, if implemented
within AM-DEGL, is well suited for selected problems only. Note that in SADE the scale factor is generated from
N(0.5, 0.3), hence 95% of its values falls within a large interval [0.1, 1.1], when in two other ancestor algorithms of
AM-DEGL (DEGL and SFMDE) scale factors around 0.70.8 are suggested.
3. The use of greedy local mutation model (LM-greedy) leads to similar, although mitigated, effect as noted for F-SADE var-
iant and is not recommended.
AM-DEGL has been selected as the best among compared novel variants because the averaged over all problems number
of algorithms that perform signicantly better than AM-DEGL is the smallest, hence this method outperform other variants
in terms of reliability. Note that other variants of AM-DEGL may be obtained easily by modifying only one of AM-DEGL ele-
ments (see pseudocode in Fig. 1).
6. Empirical results and No Free Lunch theorems for optimization
In the present paper a novel DE algorithm has been proposed and its good performance was empirically veried for 25
benchmark problems. Although this is a very frequent approach, one may ask if it is justied to propose a novel heuristic
and show its usefulness based on small number of benchmark problems if according to NFL [149,150,73] the expected per-
formance of all possible algorithms averaged over all possible problems is equal. The fact that in many papers NFL is simply
ignored or just mentioned with a brief comment has been criticized in the literature [99,73]. Because understanding the lim-
itations of the novel methods is no less important than conrming their good performance for some problems, this section is
devoted to a short literature review of NFL and its interpretations, followed by simple experiment that empirically shows the
limitations of DE algorithms. For the more detailed discussion of NFL and its proofs the reader is referred to the classical lit-
erature [149151,37,38,120,73,65,66,7,147,118]. Below the notation proposed in [150,151] is used.
6.1. No Free Lunch theorem
To prove NFL Wolpert and Macready [149151] specied that:
1. NFL aims at combinatorial optimization, where both the search space X and the space of tness values Y are nite. As
pointed out in [150], when digital computers are used these restrictions are always met in practice, even if one claims
(like in this and many other papers) that the algorithms are developed for continuous optimization of real valued func-
tions. As a result the size of the space [X[ and [Y[, as well as the size of the space of all possible problems [F[ = [Y[
[X[
is nite,
although frequently very large.
2. Each heuristic algorithm samples the elements x from X and evaluates their tness y = f(x) (x X, f(x) Y) according to the
rule that is based only on information about previously visited xs and their objective function values. NFL is valid for both
deterministic and non-deterministic algorithms [150,38]. The important, usually not fullled assumption is that an algo-
rithm does not re-visit the sampled elements. Hence NFL, while comparing algorithms, refers only to the m distinct ele-
ments as a sample d
m
= d
x
m
(1); d
y
m
(1)
_ _
; . . . ; d
x
m
(m); d
y
m
(m)
_ _ _ _
, where d
x
m
(i) is the i-th sampled distinct element within
sample of size m and d
y
m
(i) is its associated tness value. d
m
does not contain the re-visited cases. This may affect the com-
parison if algorithmin fact re-visits some elements or the problems with noise in tness are considered (e.g. F4 and F17 in
[126]).
3. An algorithm performs mapping from a sample d
m
to d
x
m1
(m 1) that is dened by the probability distribution
P
m
d
x
m1
(m 1) = x[d
m
_ _
. Hence, the algorithm a is a specication of P
i
for different values of i [150,73,151].
With the above assumptions, NFL says that the averaged over all possible problems f (f F), specied with distribution
P(f) the expected performance (measured by a performance measure U(d
y
m
)) of an algorithm a, after visiting m distinct ele-
ments x, is E(U[m; a) =
f F
U d
y
m
_ _
P d
y
m
[f ; m; a
_ _
P(f ). When
1. P(f) = 1/[F[ (NFL version proved in [150]), or when
2. P(f) is uniform over any set of functions F
s
(F
s
# F) that is closed under permutations (c.u.p.) and P(f) equal 0 if f R F
s
(so-
called Sharpened NFL proved in [120], see also [38,66,151]),
then the expected performance is independent of a, in other words the expected performance do not depend on the algo-
rithm used.
When one is interested in U d
y
m
_ _
= min
i
d
y
m
(i)
_ _
, what is frequently the case of EAs, then m < [X[ is assumed. The explana-
tion of NFL is very simple [150,24,73,70]: if one considers all possible problems (by summing over all possible fs or all prob-
lems that are c.u.p. with equal P(f)), the already visited elements do not provide any information about the unvisited ones (all
A.P. Piotrowski / Information Sciences 241 (2013) 164194 187
combinations of tness values are equally possible for unvisited elements). It is worth noting that in [66,125] NFL has also
been proved for some specic non-uniform P(f)s.
6.2. Implications, limitations and interpretations of NFL
Implications of NFL for application of heuristic optimization methods to problems that may be of any practical interest to
anyone have been studied in a number of papers from different points of view. They led to the conclusions that NFL does not
dismiss the applicability of heuristic optimization algorithms to large number of practical problems, and hence do not rule
out the need to develop novel methods.
According to [73] NFL is sometimes considered by practitioners simply as a lack of universal applicability of any heuristic
method. Based on intuition and empirical experience it is easy to agree that some algorithms perform better for some sets of
problems than the others. Although from a practical point of view this seems correct when one aims at solving a chosen
problem, this attens the NFL ndings. According to NFL the expected performance of any heuristic algorithm is the same.
As pointed out in [150] if one aims at minimization, purely random sampling, hill descending or hill climbing perform
equally well as any other heuristic. However, few people search for minimum by random sampling or using hill climbing algo-
rithm (or any other method that aims at maximization). This issue will be addressed once more in Section 6.3.
In a few papers the question whether NFL can be generalized to continuous functions was asked. In [7] it is shown that
NFL does not hold for continuous functions, only the weaker form of NFL does hold. Streeter [125] argued that uniform dis-
tributions cannot be assumed when F is innite, and that NFL would hold only when specic restrictions on P(f) are set. How-
ever, Rowe et al. [118] reinterpreted NFL using set theory approach showing that NFL does hold even for innite domains.
The rst suggestion that the differences in performance between algorithms may be found despite NFL was given in [150]
and refers to the possible head-to-head minimax distinction between two different deterministic algorithms. Although DE is
considered as a stochastic method, one must note that in [150] it is argued that when pseudorandom numbers are generated
given a seed, the algorithms become in fact deterministic. Head-to-head minimax distinction is dened as follows: if for
some minimization problems f
k
F the difference between expected performance of algorithms a and b is not lower than
k (k > 0) but does not exist any problem f
kk
F for which the difference between expected performance of algorithms b
and a is not lower than k, then algorithm b is minimax superior to algorithm a. In [150] it is empirically shown that it is pos-
sible to nd minimax distinctions between algorithms for some specied small [X[, [Y[ and some specic m and U. This means
that, although the expected performance of two different algorithms is equal for all possible problems, one algorithm may
perform much better than the other for a few problems (the difference in their performance is at least equal k), when the
other algorithm may be only marginally better, but for large number of problems. As marginal differences between perfor-
mance of different optimization methods are rarely important in practice and are frequently ruled out (for example when the
statistical tests are used to compare the results obtained by different algorithms), the lack of NFL for minimax distinction
between algorithms may be important from practical point of view. Unfortunately, in the mentioned example for which
minimax distinction is shown in [150] the overlaps between elements x visited by algorithms a and b within d
m
samples
are noted and U d
y
m
_ _
is of limited interest. Such minimax distinction are a kind of a priori distinctions between algorithms
for discussion and another example of such distinction the reader is referred to [73].
Probably the most important issue regarding NFL is that most among all problems (F) are unlikely to be of interest to any-
one [77,66]. NFL holds when all possible problems have the same P(f) [151], or when all problems within a subset of prob-
lems that are c.u.p. (F
s
) have the same P(f) [120]. As shown in [65,66], the fraction of all possible problems that are c.u.p.
converge to zero very quickly when [X[ is increasing. According to [150] when real-valued problems are implemented on dig-
ital machines, [X[ and [Y[ are very large, hence the probability that a randomly chosen subset of problems is c.u.p. converges
to zero. Nonetheless, for large [X[ and [Y[ the total number of subsets of problems that are c.u.p. is very large. The question
arise, can some relevant classes of problems be c.u.p.? In [66] it is proven that if one requires that for all f F
ss
the elements x
a
and x
b
for which f(x
a
) is a global minimum and f(x
b
) is a global maximum cannot be neighbored, then F
ss
is not c.u.p. As prob-
lems for which the global minimum and the global maximum are neighbored are not important in practice, at least because
they would not allow for robust solutions (on robust optimization see [10]) they may be excluded from consideration. In-
stead of excluding only problems for which global minima and maxima are neighbored, one may set weaker constraints
[24], for example by specifying the maximum allowed variation of f(x
n
) (when x
n
is any neighbor of x) in the neighborhood
of each x. For discussion on other classes of problems for which NFL does not hold the reader is referred to [65,125,66].
According to above discussion NFL only holds for subsets of problems that include functions that cannot be of interest to
anyone. It was also shown that some algorithms (e.g. sub-median seeker and sub-threshold-seeker) outperform some others
(e.g. random search) for some classes of functions that are not c.u.p. [24,146,70].
On the other hand, a Focused No Free Lunch (FNFL) [147] should be considered as a caution of over-enthusiastic claims
that a subsets of problems that may be of interest to anyone are not c.u.p. FNFL is based on Sharpened NFL [120] and show
that when limited number of function evaluations (m [X[) is allowed (what is frequent in EAs) the performance of a number
of specic algorithms averaged over a set of problems F
foc
, such that F
foc
F
s
# F, is equal. All problems within F
foc
may be of
practical importance. Although, contrary to NFL, the FNFL is valid only for some specic algorithms compared on specic m
and F
foc
, in practical applications m [X[ and one compares the performance of limited number of algorithms that bear some
similarities on a small number of benchmark problems that are articially constructed from a simple functions. Hence, FNFL
is at least a warning that when algorithms are compared on a small number of problems that belong to F
foc
, another small
188 A.P. Piotrowski / Information Sciences 241 (2013) 164194
number of problems that also belong to F
foc
and are important from practical point of view may exist, for which the perfor-
mance of compared algorithms would be opposite.
6.3. An alluring continuous and differentiable function
Although NFL only holds for subsets of problems among which some cannot be of interest to anyone, some of its impli-
cations suggested in [150] may help understanding the limitations of the heuristic optimization methods. Consider as an
example a very simple version of a continuous, differentiable and scalable problem, which two-dimensional version is de-
picted in Fig. 4:
y =
D
i=1
1
x
10
i
x
i
_ _
; x [0:1; 1000[
D
x
+
i
= 0:1 f (x
+
) ~ D E10 (24)
In the present sub-section all 14 DE algorithms compared in Section 5 are used to nd the minimum of 2-dimensional ver-
sion of this problem (note that for better view the illustration on Fig. 4 is cut off for x
i
> 10). When D = 2, three local minima
exists: shallow local minimum at xs = [1000, 1000] (with f(xs) = 2000) and two local minima with f(x) ~ 1E10 at
x = [0.1, 1000] and x = [1000, 0.1]. The maximum is located at x
than to xs. Maximumnumber of function calls is set to 10,000D, as for problems addressed in Section 5. A few minor
modications of control parameters of some algorithms are needed to apply them effectively for 2-dimensional problems,
because:
DEGL the neighborhood size must be large enough to include at least 2 nearest neighbors for each individual;
SFMDE population size of PSO algorithm (consult [18] for details) cannot be higher than SFMDE population size (here
NP = 5D = 10);
JADE and AdapSS-JADE the subset of p100% individuals cannot be empty;
DECLS, DEahcSPX and CoDE NP cannot be 1D as DE cannot work with just two individuals (it is increased to 5D);
In addition to DE methods, a purely random sampling is applied. The 30-run averaged results are reported in Table 6. The
mean performance of random sampling is signicantly better than the mean performance of all tested DE algorithms. Only
four DE methods were able to locate a local minimum with tness value ~ (10)
10
once (hence the averaged tness of
3.3(10)
8
). None DE method was able to converge to the global minimum. The reason of poor performance of DE methods
is simple if within initial few generations the method do not nd by chance any solution outside the basin of attraction of
xs, the whole population is quickly attracted to xs due to a concept to follow the better performing individuals, according
to which heuristics are constructed. Random sampling is not attracted anywhere, hence the possibility of nding a solution
Fig. 4. 2-Dimensional illustration of a simple differentiable problem for which the heuristic algorithms are allured to a local optimum. The problem dened
in the text is bounded by 0.1 6 x
1
, x
2
6 1000. On the left picture, on which only points with objective function value (OF) > 30. are shown, it is seen that
heuristic approaches would be allured to the local minimum at high x
1
and x
2
values. On the right picture the full span of objective function values is shown,
with global minimum at [0.1, 0.1]. For the sake of clarity both gures are cut to x
1
, x
2
< 10.
Table 6
Thirty-run averaged results for 2-dimensional variant of the proposed problem.
Random AM-DEGL DE DE-SG DEGL SADE SspDE JADE
Mean 2.8E9 2000 2000 3.3E8 2000 2000 3.3E8 2000
AdapSS-JADE DECLS DEahc-SPX CoDE EPSDE SF-MDE CDE CoDE
Mean 2000 2000 2000 2000 2000 3.3E8 3.3E8 2000
A.P. Piotrowski / Information Sciences 241 (2013) 164194 189
close to other optima by chance is the same throughout the whole run. Note that EAs like CMA-ES [57,58] would perform
even poorer than DE algorithms. This show empirically that for some problems the random sampling may indeed perform
better than heuristic optimization algorithms, as suggested in [150].
Note that although global maximum is much closer to the global minimum than the shallow minimum at xs, if the
problem dened by Eq. (24) is encoded on a computer, there is a very large number of elements between global maximum
and global minimum, which are still far frombeing neighbored (see discussion in Section 6.2). One may imagine a large num-
ber of similar problems (some very simple one-dimensional ones are given in [99]). It may seem that such problems are
rarely considered when different heuristics are compared, but this is not necessarily true. A slightly similar concept of deep
narrow global optimum on the bound was also exploited in problem F8 proposed in [126]. Can random sampling be effec-
tive for minimization of F8 problem? To address this question it was applied for 10-dimensional version of F8, which was
already used to compare DE methods in Section 5.3 (with maximum number of function calls set to 10.000D, the computa-
tions were repeated 30 times). The 30-run averaged performance of random sampling was found to be 119.63, hence it
performs poorer than most DE algorithms (see Table 2). However, random sampling performs almost equally well as SspDE
and better than DECLS and DEahcSPX. This conrms empirically that even some very popular heuristic methods used to solve
some widely known problems may perform poorer than random sampling.
Consider problem dened in Eq. (24) again. Note that even when the scope is to nd the minimum, all 14 DE algorithms
may be applied to search for the maximum of this problem. The question is what will be the lowest objective function value
found accidentally during run when DE variants are used for maximization of the problem dened by Eq. (24). This is an
analog of switching from hill descending (intuitively suggested when one aims at minimization) to hill climbing (intuitively
a wrong concept when minimization is required) as discussed in [150] this intuitive selection is not necessarily correct. The
30-run averaged lowest function values found by the way by DE algorithms during search for the maximum of 2-dimen-
sional problem dened by Eq. (24) are shown in Table 7. They are signicantly lower than values found during minimization
(reported in Table 6) for all DE variants. Interestingly, 5 DE methods used for maximization show even better minimization
performance than random sampling. Hence, for this very simple problem which minimum is to be found, heuristics that aim
at maximization outperform signicantly heuristics that aim at minimization what is not a surprising according to [150].
The results showed in this sub-section should be considered as an empirical illustration of limiting applicability of heu-
ristic methods and may serve as another precaution fromgeneralizing the good performance of any heuristic optimizer. They
may also help avoiding the over-enthusiastic opinion regarding a novel method, which may turn out pernicious for both
authors and readers (see discussion in [27]). The presented discussion also bridges the gap between different concepts in
algorithmic design which were mentioned in the Introduction, as both starting from tabula rasa following Ockhams razor
philosophy [63,19] or building on the experience following inspiration by the process of evolution may easily fail if the path
to the optimum cannot be deduced from the knowledge accumulated by the optimization algorithm during search. However,
it is hoped that AM-DEGL algorithm proposed in this paper can be successfully applicable to wide number of problems that
may be of interest, although surely not all of them.
7. Conclusions
In the present paper a novel Differential Evolution algorithm (AM-DEGL) is proposed. It is developed following an idea of
getting together the most important components of different already successful Differential Evolution methods. The novel
method puts together the slightly modied concept of adaptation of control parameters and mutation strategies proposed
in Self-Adaptive Differential Evolution [115], the concept of Global and Local mutation operators proposed in Differential
Evolution with Global and Local neighborhood-based mutation algorithm [29] and the idea of using NelderMead algorithm
as a local search method within Memetic Differential Evolution framework proposed in [18]. It is empirically shown that the
proposed method outperforms 13 various Differential Evolution algorithms on a popular set of 30- and 50-dimensional
benchmark problems proposed in [126]. The proposed method also shows reasonable performance for 10-dimensional prob-
lems. A number of variants of novel algorithm are tested and it is found that the performance of a novel approach is mod-
erately sensitive to small changes in its constituent components. For the majority of tested problems AM-DEGL is able both
to converge to good solutions quickly and to keep improving until the maximum number of function calls is reached, show-
ing a good balance between exploitation and exploration properties.
As a novel heuristic is proposed in this paper, the implications of No Free Lunch theorems for optimization are very briey
discussed. It is known that No Free Lunch hold when one considers all possible problems or a subset of problems that are
Table 7
The lowest found function values (averaged over 30-runs) of 2-dimensional variant of the proposed problem, when each DE algorithm searched for the
maximum of the function.
Random AM-DEGL DE DE-SG DEGL SADE SspDE JADE
Mean 2.8E9 1.9E9 2.1E9 4.2E9 3.1E9 2.0E9 2.4E9 3.8E9
AdapSS-JADE DECLS DEahc-SPX CoDE EPSDE SF-MDE CDE CoDE
Mean 2.9E9 1.9E9 4.0E9 2.4E9 2.6E9 2.7E9 2.5E9 2.3E9
190 A.P. Piotrowski / Information Sciences 241 (2013) 164194
closed under permutations, what is not the case for problems that may be of any practical interest. But this by no means
exclude the possibility that for many problems of interest the random sampling may perform better than heuristic algo-
rithms, what was postulated by No Free Lunch theorems [150]. A very simple continuous and differentiable problem is pro-
posed in this paper for which all considered Differential Evolution methods perform poorer than random sampling.
Moreover, all considered Differential Evolution algorithms aiming at maximization of the proposed problem found lower
objective function values than all Differential Evolution algorithms aiming at minimization. It is also found that at least
for one among 25 benchmark problems dened in [126], that are widely used to compare Evolutionary Algorithms, the ran-
dom sampling perform equally well as one, and better than two other well known Differential Evolution methods from the
literature. This should be a precaution from generalizing the good performance of any proposed heuristics.
Acknowledgments
This work has been supported by the Inner Grant of the Institute of Geophysics, Polish Academy of Sciences Nr. 1b/IGF
PAN/2012/M.
References
[1] H.A. Abbass, An evolutionary articial neural networks approach for breast cancer diagnosis, Articial Intelligence in Medicine 25 (3) (2002) 265281.
[2] J. Akhtar, B.B. Koshul, M.M. Awais, A framework for evolutionary algorithms based on Charles Sanders Peirces Evolutionary semiotics, Information
Sciences 236 (2013) 93108.
[3] R.M. Alguliev, R.M. Aliguliyev, C.A. Mehidyev, Sentence selection for generic document summarization using an adaptive differential evolution
algorithm, Swarm and Evolutionary Computation 1 (2011) 213222.
[4] M.M. Ali, Differential evolution with preferential crossover, European Journal of Operational Research 181 (3) (2007) 11371147.
[5] M.M. Ali, Differential evolution with generalized differentials, Journal of Computational and Applied Mathematics 235 (8) (2011) 22052216.
[6] J. Apolloni, G. Leguizamon, J. Garcia-Nieto, E. Alba, Island based distributed differential evolution: an experimental study on hybrid test beds, in: Proc.
of IEEE 8th International Conference on Hybrid Intelligent Systems, Barcelona, Spain, 2008, pp. 696701.
[7] A. Auger, O. Teytaud, Continuous lunches are free, in: Proc. Genet. Evolut. Comput. Conf., 2007, pp. 117129.
[8] K. Bandurski, W. Kwedlo, A lamarckian hybrid of differential evolution and conjugate gradients for neural network training, Neural Processing Letters
32 (1) (2010) 3144.
[9] R.L. Becerra, C.A. Coello, Cultured differential evolution for constrained optimization, Computer Methods in Applied Mechanics and Engineering 195
(33-36) (2006) 43034322.
[10] H.G. Beyer, B. Sendhoff, Robust optimization a comprehensive survey, Computer Methods in Applied Mechanics and Engineering 196 (3334)
(2007) 31903218.
[11] X.J. Bi, J. Xiao, Classication-based self-adaptive differential evolution with fast and reliable convergence performance, Soft Computing 15 (8) (2011)
15811599.
[12] B. Boussaid, A. Chatterjee, P. Siarry, M. Ahmed-Nacer, Two-stage update biogeography-based optimization using differential evolution algorithm
(DBBO), Computers & Operations Research 38 (2011) 11881198.
[13] J. Brest, S. Greiner, B. Boskovic, M. Mernik, V. Zumer, Self-adapting control parameters in differential evolution: a comparative study on numerical
benchmark problems, IEEE Transactions on Evolutionary Computation 10 (6) (2006) 646657.
[14] J. Brest, B. Boskovic, S. Greiner, V. Zumer, M.S. Maucec, Performance comparison of self-adaptive and adaptive differential evolution algorithms, Soft
Computing 11 (2007) 617629.
[15] J. Brest, M.S. Maucec, Population size reduction for the differential evolution algorithm, Applied Intelligence 29 (3) (2008) 228247.
[16] J. Brest, M.S. Maucec, Self-adaptive differential evolution algorithm using population size reduction and three strategies, Soft Computing 15 (11)
(2011) 21572174.
[17] Z.H. Cai, W.Y. Gong, C.X. Ling, H. Zhang, A clustering-based differential evolution for global optimization, Applied Soft Computing 11 (1) (2011) 1363
1379.
[18] A. Caponio, F. Neri, V. Tirronen, Super-t control adaptation in memetic differential evolution frameworks, Soft Computing 13 (89) (2009) 811831.
[19] F. Carafni, F. Neri, G. Iacca, A. Mol, Parallel memetic structures. Information Sciences 227 (2013) 6082.
[20] P. Chakraborty, G.G. Roy, S. Das, D. Jain, A. Abraham, An improved harmony search algorithm with differential mutation operator, Fundamenta
Informaticae 95 (4) (2009) 401426.
[21] L. Chang, C. Liao, W.B. Lin, L.L. Chen, X. Zheng, A hybrid method based on differential evolution and continuous ant colony optimization and its
application on wideband antenna design, Progress in Electromagnetics Research PIER 122 (2012) 105118.
[22] X.S. Chen, Y.S. Ong, M.H. Lim, K.C. Tan, A multi-facet survey on memetic computation, IEEE Transactions on Evolutionary Computation 15 (5) (2011)
591607.
[23] J.P. Chiou, C.F. Chang, C.T. Su, Ant direction hybrid differential evolution for solving large capacitor placement problems, IEEE Transactions on Power
Systems 19 (4) (2004) 17941800.
[24] S. Christensen, F. Oppacher, What can we learn from no free lunch? A rst attempt to characterize the concept of a searcheable function, in: Proc.
Genet. Evolut. Comput. Conf., 2001, pp. 12191226.
[25] L.D. Coelho, D.L.D. Bernert, A modied ant colony optimization algorithm based on differential evolution for chaotic synchronization, Expert Systems
with Applications 37 (6) (2010) 41984203.
[26] A. Cong, W.X. Cong, Y.J. Lu, P. Santago, A. Chattzioannou, G. Wang, Differential evolution approach for regularized bioluminescence tomography, IEEE
Transactions on Biomedical Engineering 57 (9) (2010) 22292238.
[27] M. Crepinsek, S.H. Liu, L. Mernik, A note on teaching-learning-based optimization algorithm, Information Sciences 212 (2012) 7993.
[28] S. Das, A. Konar, U.K. Chakraborty, Improving particle swarm optimization with differentially perturbed velocity, in: Proc. Genet. Evol. Comput. Conf.
(GECCO), 2005, pp. 177184.
[29] S. Das, A. Abraham, U.K. Chakraboty, A. Konar, Differential evolution using a neighborhood-based mutation operator, IEEE Transactions on
Evolutionary Computation 13 (3) (2009) 526553.
[30] S. Das, S. Sil, Kernel-induced fuzzy clustering of image pixels with an improved differential evolution algorithm, Information Sciences 180 (8) (2010)
12371256.
[31] S. Das, P.N. Suganthan, Differential evolution: a survey of the state-of-the-art, IEEE Transactions on Evolutionary Computation 15 (1) (2011) 2754.
[32] S. Das, S. Maity, B.Y. Qu, P.N. Suganthan, Real-parameter evolutionary multimodal optimization a survey of the state-of-the-art, Swarm and
Evolutionary Computation 1 (2011) 7188.
[33] I. De Falco, A. Della Cioppa, D. Maisto, U. Scafuri, E. Tarantino, Biological invasioninspired migration in distributed evolutionary algorithms,
Information Sciences 207 (2012) 5065.
A.P. Piotrowski / Information Sciences 241 (2013) 164194 191
[34] V.V. De Melo, A.C.B. Delbem, Investigating smart sampling as a population initialization method for differential evolution in continuous problems,
Information Sciences 193 (2012) 3653.
[35] J. Demsar, Statistical comparisons of classiers over multiple data sets, Journal of Machine Learning Research 7 (2006) 130.
[36] B. Dorronosoro, P. Bouvry, Improving classical and decentralized differential evolution with new mutation operator and population topologies, IEEE
Transactions on Evolutionary Computation 15 (1) (2011) 6798.
[37] S. Droste, T. Jansen, I. Wegener, Perheps not a Free Lunch but at Least a Free Appetizer, Tech. Rep. ISSN 1433-3325, Dep. Comp. Sci. University of
Dortmund, 1998.
[38] S. Droste, T. Jansen, I. Wegener, Optimization with randomized search heuristics the (A)NFL theorem, realistic scenarios, and difcult functions,
Theoretical Computer Science 287 (1) (2002) 131144.
[39] A.E. Eiben, R. Hinterding, Z. Michalewicz, Parameter control in evolutionary algorithms, IEEE Transactions on Evolutionary Computation 3 (2) (1999)
124141.
[40] A.E. Eiben, S.K. Smit, Parameter tuning for conguring and analyzing evolutionary algorithms, Swarm and Evolutionary Computation 1 (2011) 1931.
[41] M.G. Epitropakis, V.P. Plagianakos, M.N. Vrahatis, Balancing the exploration and exploitation capabilities of the differential evolution algorithm, in:
Proc. of the IEEE Congress on Evolutionary Computation, Hong Kong, China, 2008, pp. 26862693.
[42] M.G. Epitropakis, V.P. Plagianakos, M.N. Vrahatis, Evolutionary adaptation of the differential evolution control parameters, in: Proc. of the IEEE
Congress on Evolutionary Computation, Trondheim, Norway, 2009. pp. 13591366.
[43] M.G. Epitropakis, D.K. Tasoulis, N.G. Pavlidis, V.P. Plagianakos, M.N. Vrahatis, Enhancing differential evolution utilizing proximity-based mutation
operations, IEEE Transactions on Evolutionary Computation 15 (1) (2011) 99119.
[44] M.G. Epitropakis, V.P. Plagianakos, M.N. Vrahatis, Evolving cognitive and social experience in particle swarm optimization through differential
evolution: a hybrid approach, Information Sciences 216 (2012) 5092.
[45] H.Y. Fan, J. Lampinen, A Trigonometric mutation operation to Differential Evolution, Journal of Global Optimization 27 (1) (2003) 105129.
[46] V. Feoktisov, Differential evolution in search of solutions, Book Series: Springer Optimization and Its Applications, vol. 5, Springer, New York, 2006.
[47] M. Friedman, The use of ranks to avoid the assumption of normality implicit the in the analysis of variance, Journal of American Statistical Association
32 (1937) 674701.
[48] R. Gamperle, S.D. Muller, P. Koumoutsakos, A parameter study of differential evolution, in: Advances in Intelligent Systems, Fuzzy Systems,
Evolutionary Computation, WSEAS Press, Interlaken, Switzerland, 2002.
[49] S. Garcia, A. Fernandez, J. Luengo, F. Herdera, Advanced nonparametric tests for multiple comparison in the design of experiments in computational
intelligence and data mining: experimental analysis of power, Information Sciences 180 (10) (2010) 20442064.
[50] A. Ghosh, S. Das, A. Chowdhury, R. Gini, An improved differential evolution algorithm with tness-based adaptation of the control parameters,
Information Sciences 181 (18) (2011) 37493765.
[51] S. Ghosh, S. Das, S. Roy, S.K.M. Islam, P.N. Suganthan, A differential covariance matrix adaptation evolutionary algorithm for real parameter
optimization, Information Sciences 181 (1) (2012) 199219.
[52] W.Y. Gong, Z.H. Cai, C.X. Ling, H. Li, Enhanced differential evolution with adaptive strategies for numerical optimization, IEEE Transactions on
Systems, Man and Cybernetics, Part B Cybernetics 41 (2) (2011) 397413.
[53] W.Y. Gong, Z.H. Cai, C.Y. Ling, DE/BBO: a hybrid differential evolution with biogeography-based optimization for global numerical optimization, Soft
Computing 15 (4) (2011) 645665.
[54] W.Y. Gong, Z.H. Cai, Differential evolution with ranking-based mutation operators, IEEE Transactions on Systems, Man and Cybernetics, Part B
Cybernetics, http://dx.doi.org/10.1109/TCYB.2013.2239988.
[55] W.Y. Gong, A. Fialho, Z.H. Cai, H. Li, Adaptive strategy selection in differential evolution for numerical optimization: an empirical study, Information
Sciences 181 (24) (2011) 53645386.
[56] N. Hachicha, B. Jarboui, P. Siarry, A fuzzy logic control using a differential evolution algorithm aimed at modelling the nancial market dynamics,
Information Sciences 181 (1) (2011) 7991.
[57] N. Hansen, A. Ostermeier, Adapting arbitrary normal mutation distributions in evolution strategies: the covariance matrix adaptation, in: Proc. IEEE
Int. Conf. Evol. Comput., Nagoya, Japan, 1996, pp. 312317.
[58] N. Hansen, A.S.P. Niederberger, L. Guzzella, P. Koumoutsakos, A method for handling uncertainty in evolutionary optimization with an application to
feedback control of combustion, IEEE Transactions on Evolutionary Computation 13 (1) (2009) 180197.
[59] T. Hendtlass, A combined swarm differential evolution algorithm for optimization problems, in: Lecture Notes in Computer Science, vol. 2070, 2001,
pp.1118.
[60] S. Holm, A simple sequentially rejective multiple test procedure, Scandinavian Journal of Statistics 6 (1979) 6570.
[61] X. Hu, T. Glenn, F. Scalzo, M. Bergsneider, C. Sarkiss, N. Martin, P. Vespa, Intracranial pressure pulse morphological features improved detection of
decreased cerebral blood ow, Physiological Measurement 31 (5) (2010) 679695.
[62] F.Z. Huang, L. Wang, Q. He, An effective co-evolutionary differential evolution for constrained optimization, Applied Mathematics and Computation
186 (1) (2007) 340356.
[63] G. Iacca, F. Neri, E. Mininno, Y.S. Ong, M.H. Lim, Ockhams razor in memetic computing: three stage optimal memetic exploration, Information
Sciences 188 (2012) 1743.
[64] G. Iacca, F. Neri, E. Mininno, Noise analysis compact differential evolution, International Journal of Systems Science 43 (7) (2012) 12481267.
[65] C. Igel, M. Toussaint, On classes of functions for which no free lunch results hold, Information Processing Letters 86 (6) (2003) 317321.
[66] C. Igel, M. Toussaint, A no-free lunch theorem for non-uniform distributions of target functions, Journal of Mathematical Modelling and Algorithms 3
(2004) 313322.
[67] S.M. Islam, S. Das, S. Ghosh, S. Roy, P.N. Suganthan, An adaptive differential evolution algorithm with novel mutation and crossover strategies for
global numerical optimization, IEEE Transactions on Systems, Man and Cybernetics, Part B Cybernetics 42 (2) (2012) 482500.
[68] G. Jia, Y. Wang, Z. Cai, Y. Jin, An improved (l k)-constrained differential evolution for constrained optimization, Information Sciences 222 (2013)
302322.
[69] D.L. Jia, G.X. Zheng, M.K. Khan, An effective memetic differential evolution algorithm based on chaotic local search, Information Sciences 181 (15)
(2011) 31753187.
[70] P. Jiang, Y.P. Chen, Free lunches on the discrete Lipschitz class, Theoretical Computer Science 412 (17) (2011) 16141628.
[71] P. Kaelo, M.M. Ali, A numerical study on some modied differential evolution algorithms, European Journal of Operational Research 169 (3) (2006)
11761184.
[72] T.G. Kolda, R.M. Lewis, V. Torczon, Optimization by direct search: new perspectives on some classical and modern methods, SIAM Review 45 (3)
(2003) 385482.
[73] M. Koppen, D.H. Wolpert, W.G. Macready, Remarks on a recent paper on the No free lunch theorems, IEEE Transactions on Evolutionary
Computation 5 (3) (2001) 295296.
[74] N. Krasnogor, J. Smith, A tutorial for competent memetic algorithms: model, taxonomy, and design issues, IEEE Transactions on Evolutionary
Computation 9 (5) (2005) 474488.
[75] J.C. Lagarias, J.A. Reeds, M.H. Wright, P.E. Wright, Convergence properties of the NelderMead simplex method in low dimensions, SIAM Journal of
Optimization 9 (1) (1998) 112147.
[76] J. Lampinen, I. Zelinka, On stagnation of the differential evolution algorithm, in: Proceedings of 6th International Mendel Conference on Soft
Computing, Brno, Czech Republic, 2000.
192 A.P. Piotrowski / Information Sciences 241 (2013) 164194
[77] W.B. Langdon, R. Poli, Evolving problems to learn about particle swarm optimizers and other search algorithms, IEEE Transactions on Evolutionary
Computation 11 (5) (2007) 561578.
[78] T.W. Liao, Two hybrid differential evolution algorithms for engineering design optimization, Applied Soft Computing 10 (4) (2010) 11881199.
[79] C. Lin, A.Y. Qing, Q.Y. Feng, A comparative study of crossover in differential evolution, Journal of Heuristics 17 (6) (2011) 675703.
[80] J. Liu, J. Lampinen, A fuzzy adaptive differential evolution algorithm, Soft Computing 9 (2005) 448462.
[81] H. Liu, Z. Cai, Y. Wang, Hybridizing particle swarm optimization with differential evolution for constrained numerical and engineering optimization,
Applied Soft Computing 10 (2) (2010) 629640.
[82] G. Liu, Y. Li, X. Nie, H. Zheng, A novel clustering-based differential evolution with 2 multi-parent crossovers for global optimization, Applied Soft
Computing 12 (2) (2012) 663681.
[83] R. Mallipeddi, P.N. Suganthan, Ensemble of constraint handling techniques, IEEE Transactions on Evolutionary Computation 14 (4) (2010) 561579.
[84] R. Mallipeddi, J.P. Lie, P.N. Suganthan, S.G. Razul, C.M.S. See, A differential evolution approach for robust adaptive beamforming based on joint
estimation of look direction and array geometry, Progress in Electromagnetics Research PIER 119 (2011) 381394.
[85] R. Mallipeddi, P.N. Suganthan, Q.K. Pan, M.F. Tasgetiren, Differential evolution algorithm with ensemble of parameters and mutation strategies,
Applied Soft Computing 11 (2) (2011) 16791696.
[86] A.D. Masegosa, D.A. Pelta, J.L. Verdegay, A centralised cooperative strategy for continuous optimisation: the inuence of cooperation in performance
and behaviour, Information Sciences 219 (2013) 7392.
[87] E. Mininno, F. Neri, F. Cupertino, D. Naso, Compact differential evolution, IEEE Transactions on Evolutionary Computation 15 (1) (2011) 3254.
[88] A.W. Mohamed, H.Z. Sabry, Constrained optimization based on modied differential evolution algorithm, Information Sciences 194 (2012) 171208.
[89] C.G. Moles, J.R. Banga, K. Keller, Solving nonconvex climate control problems: pitfalls and algorithm performances, Applied Soft Computing 5 (1)
(2004) 3544.
[90] P.W. Moore, G.K. Venayagamoorthy, Evolving digital circuits using hybrid particle swarm optimization and differential evolution, International
Journal of Neural Systems 16 (3) (2006) 163177.
[91] J.A. Nelder, R. Mead, A simplex-method for function minimization, Computer Journal 7 (4) (1965) 308313.
[92] F. Neri, V. Tirronen, On memetic differential evolution frameworks: a study of advantages and limitations in hybridization, in: Proc. of IEEE World
Congress on Computational Intelligence, Hong Kong, 2008, pp. 21352142.
[93] F. Neri, E. Mininno, Memetic compact differential evolution for Cartesian robot control, IEEE Computational Intelligence Magazine 5 (2) (2010) 5465.
[94] F. Neri, V. Tirronen, Recent advances in differential evolution: a survey and experimental analysis, Articial Intelligence Review 33 (12) (2010) 61
106.
[95] F. Neri, G. Iacca, E. Mininno, Disturbed exploitation compact differential evolution for limited memory optimization problems, Information Sciences
181 (12) (2011) 24692487.
[96] F. Neri, C. Cotta, Memetic algorithms and memetic computing optimization: a literature review, Swarm and Evolutionary Computation 2 (2012) 114.
[97] N. Noman, H. Iba, Inferring gene regulatory networks using differential evolution with local search heuristics, IEEE ACM Transactions on
Computational Biology and Bioinformatics 4 (4) (2007) 634647.
[98] N. Noman, H. Iba, Accelerating differential evolution using an adaptive local search, IEEE Transactions on Evolutionary Computation 12 (1) (2008)
107125.
[99] M. Oltean, Searching for a practical evidence for the no free lunch theorems, in: Bioinspired Approaches to Advanced Information Technology,
Springer, Lausanne, Switzerland, 2004.
[100] M.G.H. Omran, A. Salman, A.P. Engelbrecht, Self-adaptive differential evolution, in: Lecture Notes in Articial Intelligence, Springer-Verlag, Berlin,
Germany, 2005.
[101] M.G.H. Omran, A.P. Engelbrecht, A. Salman, Bare bones differential evolution, European Journal of Operational Research 196 (1) (2009) 128139.
[102] Y.S. Ong, A.J. Keane, Meta-lamarckian learning in memetic algorithms, IEEE Transactions on Evolutionary Computation 8 (2) (2004) 99110.
[103] Y.S. Ong, M.H. Lim, N. Zhu, K.W. Wong, Classication of adaptive memetic algorithms: a comparative study, IEEE Transactions on Systems, Man and
Cybernetics, Part B Cybernetics 36 (1) (2006) 141152.
[104] Y.S. Ong, M.H. Lim, X. Chen, Memetic computation past, present and future, IEEE Computational Intelligence Magazine 5 (2) (2010) 2431.
[105] Q.K. Pan, P.N. Suganthan, L. Wang, L. Gao, R. Mallipeddi, A differential evolution algorithm with self-adapting strategy and control parameters,
Computers & Operations Research 38 (2011) 394408.
[106] K.M. Passino, Biomimicry of bacterial foraging for distributed optimization and control, IEEE Control Systems Magazine 22 (3) (2002) 5267.
[107] F. Peng, K. Tang, G.L. Chen, X. Yao, Population-based algorithm portfolios for numerical optimization, IEEE Transactions on Evolutionary Computation
14 (5) (2010) 782800.
[108] A.P. Piotrowski, J.J. Napiorkowski, Grouping differential evolution algorithm for multi-dimensional optimization problems, Control and Cybernetics
39 (2) (2010) 527550.
[109] A.P. Piotrowski, J.J. Napiorkowski, A. Kiczko, Differential Evolution algorithm with separated groups for multi-dimensional optimization problems,
European Journal of Operational Research 216 (2012) 3346.
[110] A.P. Piotrowski, J.J. Napiorkowski, A. Kiczko, Corrigendum to: Differential evolution algorithm with separated groups for multi-dimensional
optimization problems [Eur. J. Oper. Res. 216 (2012) 3346], European Journal of Operational Research 219 (2012) 488.
[111] A.P. Piotrowski, P.M. Rowinski, J.J. Napiorkowski, Comparison of evolutionary computation techniques for noise injected neural network training to
estimate longitudinal dispersion coefcients in rivers, Expert Systems with Applications 39 (1) (2012) 13541361.
[112] K.V. Price, An introduction to differential evolution, in: New Ideas in Optimization, McGraw-Hill, London, UK, 1999.
[113] K.V. Price, R.M. Storn, J.A. Lampinen, Differential Evolution, A Practical Approach to Global Optimization, Springer-Verlag, Berlin, Heidelberg, 2005.
[114] A.K. Qin, P.N. Suganthan, Self-adaptive differential evolution algorithm for numerical optimization, in: Proc. IEEE Congress on Evolutionary
Computation, Edinburgh, UK, 2005.
[115] A.K. Qin, V.L. Huang, P.N. Suganthan, Differential evolution algorithm with strategy adaptation for global numerical optimization, IEEE Transactions
on Evolutionary Computation 13 (2) (2009) 398417.
[116] S. Rahnamayan, H.R. Tizhoosh, M.M.A. Salama, Opposition-based differential evolution, IEEE Transactions on Evolutionary Computation 12 (1) (2008)
6479.
[117] H.H. Rosenbrock, An automated method for nding the greatest or least value of a function, Computer Journal 3 (3) (1960) 175184.
[118] J.E. Rowe, M.D. Vose, A.H. Wright, Reinterpreting no free lunch, Evolutionary Computation 17 (1) (2009) 117129.
[119] R. Salomon, Re-evaluating genetic algorithm performance under coordinate rotation on benchmark functions. A survey of some theoretical and
practical aspects of genetic algorithms, BioSystems 39 (1996) 263278.
[120] C. Schumacher, M.D. Vose, L.D. Whitley, The no free lunch and problem description length, in: Proc. Genet. Evolut. Comput. Conf., 2001, pp. 565570.
[121] S. Sivananaithaperumal, J. Miruna Joe Amali, S. Baskar, P.N. Suganthsn, Constrained self adaptive differential evolution based design of robust optimal
xed structure controller, Engineering Applications of Articial Intelligence 24 (6) (2011) 10841093.
[122] J.D. Sosin, Ausonius Juvenal and the Winstedt fragment, Classical Philology 95 (2) (2000) 199206.
[123] R. Storn, K.V. Price, Differential Evolution A Simple and Efcient Adaptive Scheme for Global Optimization Over Continuous Spaces, Tech. Report TR-
95-012, International Computer Sciences Institute, Berkeley, California, USA, 1995.
[124] R. Storn, K.V. Price, Differential evolution a simple and efcient heuristic for global optimization over continuous spaces, Journal of Global
Optimization 11 (4) (1997) 341359.
[125] M.J. Streeter, Two broad classes of functions for which a no free lunch result does not hold, in: Proc. Genet. Evolut. Comput. Conf., 2003, pp. 1418
1430.
A.P. Piotrowski / Information Sciences 241 (2013) 164194 193
[126] P.N. Suganthan, N. Hansen, J.J. Liang, K. Deb, Y.P. Chen, A. Auger, S. Tiwari, Problem Denitions and Evaluation Criteria for the CEC 2005 Special
Session on Real-Parameter Optimization, Nanyang Technol. Univ., Singapore, Tech. Rep. KanGAL #2005005, IIT Kanpur, India, 2005.
[127] R. Tanese, Distributed genetic algorithms, in: Proc. of 3rd Int. Conference on Genetic Algorithms, San Francisco, CA, USA, 1989, pp. 434439.
[128] D.K. Tasoulis, N.G. Pavlidis, V.P. Plagianakos, M.N. Vrahatis, Parallel differential evolution, in: Proceedings of the IEEE Congress on Evolutionary
Computation, Portland, Oregon, USA, 2004, pp. 20232029.
[129] N.S. Teng, J. Teo, M.H.A. Hijazi, Self-adaptive population sizing for a tune-free differential evolution, Soft Computing 13 (7) (2009) 709724.
[130] J. Teo, Exploring dynamic self-adaptive populations in differential evolution, Soft Computing 11 (8) (2006) 673686.
[131] D. Thierens, An adaptive pursuit strategy for allocating operator probabilities, in: Proc. Genet. Evol. Comput. Conf., 2005, pp. 15391546.
[132] V. Tirronen, F. Neri, T. Karkkainen, K. Majava, T. Rossi, An enhanced memetic differential evolution in lter design for defect detection in paper
production, Evolutionary Computation 16 (4) (2008) 529555.
[133] V. Tirronen, F. Neri, Differential evolution with tness diversity self-adaptation, Nature-Inspired Algorithms for Optimisation, SCI 193, Springer-
Verlag, Berlin, Heidelberg, 2009, pp. 199234.
[134] I. Triguero, S. Garcia, F. Herrera, Differential evolution for optimizing the positioning of prototypes in nearest neighbor classication, Pattern
Recognition 44 (4) (2011) 901916.
[135] O. Urfalioglu, O. Arikan, Self-adaptive randomized and rank-based differential evolution for multimodal problems, Journal of Global Optimization 51
(4) (2011) 607640.
[136] K. Vaisakh, P. Praveena, S. Rama Mohana Rao, Kala Maeh, Solving dynamic economic dispatch problem with security constraints using bacterial
foraging PSO-DE algorithm, Electrical Power and Energy Systems 39 (2012) 5667.
[137] C. Voglis, K.E. Parsopoulos, D.G. Papageorgiou, I.E. Lagaris, M.N. Vrahatis, MEMPSODE: a global optimization software based on hybridization of
population-based algorithms and local searches, Computer Physics Communications 183 (5) (2012) 11391154.
[138] J.A. Vrugt, B.A. Robinson, J.M. Hyman, Self-adaptive multimethod search for global optimization in real-parameter spaces, IEEE Transactions on
Evolutionary Computation 13 (2) (2009) 243259.
[139] Y. Wang, Z. Cai, Q. Zhang, Differential evolution with composite trial vector generation strategies and control parameters, IEEE Transactions on
Evolutionary Computation 15 (1) (2011) 5566.
[140] L. Wang, L.P. Li, A coevolutionary differential evolution with harmony search for reliabilityredundancy optimization, Expert Systems with
Applications 39 (5) (2012) 52715278.
[141] Y. Wang, Z. Cai, Q. Zhang, Enhancing the search ability of differential evolution through orthogonal crossover, Information Sciences 185 (1) (2012)
153177.
[142] M. Weber, F. Neri, V. Tirronen, Distributed differential evolution with explorativeexploitative population families, Genetic Programming and
Evolvable Machines 10 (4) (2009) 343371.
[143] M. Weber, V. Tirronen, F. Neri, Scale factor inheritance mechanism in distributed differential evolution, Soft Computing 14 (11) (2010) 11871207.
[144] M. Weber, F. Neri, V. Tirronen, A study on scale factor in distributed differential evolution, Information Sciences 181 (12) (2011) 24882511.
[145] M. Weber, F. Neri, V. Tirronen, Shufe or update parallel differential evolution for large-scale optimization, Soft Computing 15 (2011) 20892107.
[146] D. Whitley, J. Rowe, Subthreshold-seeking local search, Theoretical Computer Science 361 (1) (2006) 217.
[147] D. Whitley, J. Rowe, Focused no free lunch theorems, in: Proc. Genet. Evol. Comput. Conf., 2008, pp. 811818.
[148] F. Wilcoxon, Individual comparisons by ranking methods, Biometrics Bulletin 1 (6) (1945) 8083.
[149] D.H. Wolpert, W.G. Macready, No Free Lunch Theorems for Search, Tech. Rep. SFI-TR-05-010, Santa Fe, NM, USA, 1995.
[150] D.H. Wolpert, W.G. Macready, No free lunch theorems for optimization, IEEE Transactions on Evolutionary Computation 1 (1) (1997) 6782.
[151] D.H. Wolpert, W.G. Macready, Coevolutionary free lunches, IEEE Transactions on Evolutionary Computation 9 (6) (2005) 721735.
[152] B. Xin, J. Chen, J. Zhang, H. Fang, Z.H. Peng, Hybridizing differential evolution and particle swarm optimization to design powerful optimizers: a
review and taxonomy, IEEE Transactions on Systems, Mac and Cybernetics, Part C Applications and Reviews 42 (5) (2012) 744767.
[153] D. Zaharie, Inuence of crossover on the behavior of differential evolution algorithms, Applied Soft Computing 9 (3) (2009) 11261138.
[154] T. Zhang, L. Chen, Z.H. Yan, B. Li, Design of dual offset shaped reector antenna based on DEGL algorithm, Journal of Electromagnetic Waves and
Applications 25 (56) (2011) 723732.
[155] M. Zhang, W. Luo, X.F. Wang, Differential evolution with dynamic stochastic selection for constrained optimization, Information Sciences 178 (15)
(2008) 30433074.
[156] J. Zhang, A.C. Sanderson, JADE: adaptive differential evolution with optional external archive, IEEE Transactions on Evolutionary Computation 13 (5)
(2009) 945958.
[157] S.Z. Zhao, P.N. Suganthan, S. Das, Self-adaptive differential evolution with multi-trajectory search for large-scale optimization, Soft Computing 15 (11)
(2011) 21752185.
[158] Z. Zhao, J.Z. Wang, J. Zhao, Z.Y. Su, Using a grey model optimized by differential evolution algorithm to forecast the per capita annual net income of
rural households in China, OMEGA International Journal of Management Science 40 (5) (2012) 525532.
[159] S.Z. Zhao, P.N. Suganthan, Empirical investigations into the exponential crossover of differential evolutions, Swarm and Evolutionary Computation 9
(2013) 2736.
[160] W. Zhu, Y. Tang, J.A. Fang, W. Zhang, Adaptive population tuning scheme for differential evolution, Information Sciences 223 (2012) 164191.
194 A.P. Piotrowski / Information Sciences 241 (2013) 164194