Journal of Robotics, Networking and Artificial Life

Volume 5, Issue 3, December 2018, Pages 165 - 168

Integrated Optimization of Differential Evolution with Grasshopper Optimization Algorithm

Authors
Duangjai Jitkongchuen*, Udomlux Ampant
College of Innovative Technology and Engineering, Dhurakij Pundit University, Thailand
*Corresponding author. Email: duangjai.jit@dpu.ac.th
Corresponding Author
Duangjai Jitkongchuen
Received 24 March 2018, Accepted 6 June 2018, Available Online 1 December 2018.
DOI
10.2991/jrnal.2018.5.3.5How to use a DOI?
Keywords
Meta-heuristic; differential evolution algorithm; grasshopper optimization algorithm; optimization
Abstract

This paper proposes a scheme to improve the differential evolution (DE) algorithm performance with integrated the grasshopper optimization algorithm (GOA). The grasshopper optimization algorithm mimics the behavior of grasshopper. The characteristic of grasshoppers is slow movement in the larval stage but sudden movement in the adulthood which seem as exploration and exploitation. The grasshopper optimization algorithm concept is added to DE to guide the search process for potential solutions. The efficiency of the DE/GOA is validated by testing on unimodal and multimodal benchmarks optimization problems. The results prove that the DE/GOA algorithm is competitive compared to the other meta-heuristic algorithms.

Copyright
© 2018 The Authors. Published by Atlantis Press SARL.
Open Access
This is an open access article under the CC BY-NC license (http://creativecommons.org/licences/by-nc/4.0/).

1. INTRODUCTION

Meta-heuristic algorithm techniques constitute range from simple local search procedures to complex learning processes. The basic concepts of meta-heuristics permit an abstract level description. Nature-inspired meta-heuristic can be grouped in three main classes: Evolution-based, physics-based and swarm-based.

Evolution-based algorithms are inspired by the concepts of nature evolution. The population in search process are randomly generated which are developed over subsequent generation. The next generation is created by combination of the individuals in the previous generation. These provide the population to be optimized over the course of generations. The most popular evolution-based algorithm is genetic algorithms (GA) [1]. GA has a process of fitness-based selection and recombination to produce a successor population for the next generation. During recombination, child chromosomes are produced by recombined genetic material from parent chromosomes are selected. As this process is repeated, a sequence of consecutive generations evolves and the average fitness of the chromosomes inclines to increase until some stopping criterion is reached. Other popular algorithms such as differential evolution (DE) algorithms [2], the principal difference between genetic algorithms and differential evolution is that DE use mutation as the primary search mechanism, while GA use crossover as the probabilistic mechanism to exchange of information among solutions to locate better solutions. Biogeography-based optimizer (BBO) [3] algorithm is applied from nature phenomenon regarding the distribution of living creatures in various islands. In this field of study different ecosystems are explored for finding the relations between different species in terms of immigration, emigration, and mutation. The main inspiration of the BBO algorithm is to reach a stable situation based on migration and mutation in the evolution of ecosystems.

The second class of meta-heuristic is physics-based algorithms that imitate the physical rules. For example, gravitational search algorithm (GSA) [4], GSA is a nature-inspired conceptual framework based on gravitational kinematics, a branch of physics that models the motion of masses moving under the influence of gravity. Galaxy-based search algorithm [5] is inspired on the spiral arm of spiral galaxies to search its surrounding. This spiral movement is improved by chaos to escape from local optimums. Curved space optimization [6] inspired by general relativity theory is used to enhance the efficiency of a simple random search, and convert it to a very robust optimization tool.

The third class of meta-heuristic is swarm-based algorithms that imitate the social behavior of swarms, herds, flocks, or schools of animals. The most popular algorithms are particle swarm optimization (PSO) [7]. PSO inspired by bird flocking or fish schooling. Each particle will be iteratively updated to improve its solution during movement around the search space. The result leads to the best value of global solution. Other popular swarm-based algorithms are bat algorithms (BA) [8]. BA is inspired by echolocation behavior of micro­bats to avoid obstacles, find their home, and detect their prey in the darkness. Firefly algorithm (FA) [9] applied from the flashing light of fireflies when they want to mate or to warn others about predators.

The new meta-heuristic has been focused on this paper is differential evolution algorithm which is a global search optimization algorithm and improved performance of it by added grasshopper optimization algorithm (GOA) to increase convergence rate.

The rest of the paper is structured as follows. Section 2 describes some backgrounds of the differential evolution algorithm. In Section 3, the proposed algorithm is presented. Section 4 shows the experimental results and the conclusions will be discussed in Section 5.

2. THE DIFFERENTIAL EVOLUTION ALGORITHM

The differential evolution algorithm optimizes a problem by sustaining a population of candidate solutions and creating new candidate solutions by combining existing ones according to its simple procedure, and then keeping whichever candidate solution has the best score or fitness on the optimization problem at hand.

The procedure of differential evolution algorithm starts to initialize target vectors XiG=(x1,i,x2,i,...,xj,i,...,xN,i) where i = 1, 2, ..., NP. NP is the population size. N is the dimension of the population. The superscript G identifies the Gth generation.

The target vectors are used to generate the mutant vectors ViG=(v1,i,v2,i,...,vj,i,...,vN,i) in next steps. The DE/current-to-best/1 mutation scheme is shown in Eq. (1).

ViG=XiG+F(XbestGXiG)+F(Xr1GXr2G)
where r1 and r2 are integer number chosen from the set {1, 2, ..., NP} and must be different from index i. XbestG is the best solution in the generation G. F is used to control amplification of the differential evolution. The value of F is set in the range [0, 2]. The target vectors and mutant vectors are used to generate the trial vectors UiG=(u1,i,u2,i,...,uj,i,...,uN,i) in crossover operation to increase the diversity of population. The trial vectors are generated by Eq. (2).
uj,i={vj,i,if(randCR)or(j=jrand)xj,i,otherwise
where j = 1, 2, ..., N. rand is a uniform random number chosen in the range [0, 1]. CR ∈ [0, 1] is the crossover parameter. jrand is an index randomly chosen in the range [0, N].

The last step is the selection operation to choose a better vector for next generation. The new generations are selected by comparing the fitness value between the target vector and the trial vector. The selection is shown in Eq. (3).

XiG+1={UiG,if(f(UiG)<f(XiG))XiG,otherwise
The whole process is repeated until the termination criteria are satisfied or a predefined number of iterations are reached.

3. THE PROPOSED ALGORITHM

The schema of DE/current-to-best/1 in mutation operation is updated by splitting F mutation parameter to λ and F. The new mutation schema is shown in Eq. (4).

ViG=XiG+λ(XbestGXiG)+F(Xr1GXr2G)
where λ is used to balance the difference vectors between the best vectors and the target vectors. The value of λ that is generated by Eq. (5) is decreased in each generation.
λG+1=λG(λG×c1)
Meanwhile the mutation parameter F is used to perform the amplification of the difference between two random population members. When the procedure is trapped in the local optimum, the value of F is increased by Eq. (6).
FG+1=FG+(FG×c2)
The parameters c1 in Eq. (5) and c2 in Eq. (6) are uniform random numbers. c1 ∈ [0, 0.1] is a random value and c2 ∈ [0, 1] is a random value. The crossover parameter (CR) is regenerated at the end of each generation to find an optimized parameter. The new crossover parameter is generated by Eq. (7).
CRG+1=CRG+(meanA(SCR)×c3)
where meanA is the usual arithmetic mean. SCR is the set of crossover parameter values. C3 ∈ [0, 1] is a uniform random number.

After that, the crossover operation is rearranged. The vectors that created from the target vectors and the mutant vector is called the prime vectors (U′i). The prime vectors is used to generate the trial vectors by comparing the fitness value between the prime vectors and the mutant vectors, otherwise the trial vectors is generated by grasshopper optimization algorithm.

The grasshopper optimization algorithm was proposed in Saremi et al. [10] mimics the behavior of grasshopper. The characteristic of the swarm in the larval stage is slow movement and small steps but is long range and suddenly movement in adulthood. These two characters seem as exploration and exploitation. The search agents are encouraged to move suddenly in exploration and conduce to move locally during exploitation. These two functions as well as target seeking are performed by grasshopper naturally.

The proposed algorithm takes the feature of movement of the grasshopper to generate a new value for the trial vectors is showed in Eq. (8).

Uid=c(j=1jiNPcubdlbd2s(|xjdxid|)xjxidi,j)+T^d
where ubd and lbd are the upper bound and the lower bound in the dth dimension. s(r)= fer/ler, di,j is the distance between ith and jth grasshopper. T^d is the value of the dth dimension in the target. The parameter c is decreased equivalent to the number of iteration to balancing exploration and exploitation is calculated as Eq. (9).
c=cmaxlcmaxcminL
where cmax and cmin are the maximum and the minimum value, respectively. l is the current iteration and L is the maximum number of iterations.

4. THE EXPERIMENTAL RESULTS

The proposed algorithm has been evaluated performance with thirteen benchmark functions [11]. The test functions are unimodal (f1f7) and multimodal (f8f13). The details of benchmark functions are shown in Table 1. The mutation parameters set λ, F = 0.5, the crossover parameter set CR = 0.9 and population size set NP = 100.

Functions Range
f1(x)=i=1Nxi2 [−100, 100]
f2(x)=i=1N|xi|+i=1N|xi| [−100, 100]
f3(x)=i=1N(j=1ixj)2 [−10, 10]
f4(x) = maxi {|xi|, 1 ≤ iN} [−100, 100]
f5(x)=i=1N1[100(xi+1xi2)2+(xi1)2] [−100, 100]
f6(x)=i=1N(xi+0.5)2 [−30, 30]
f7(x)=i=1Nixi4+random[0,1) [−100, 100]
f8(x)=i=1Nxisin(|xi|) [−1.28, 1.28]
f9(x)=i=1N[xi210cos(2πxi)+10] [−500, 500]
f10(x)=20exp(0.21Ni=1Nxi2)exp(1Ni=1Ncos2πxi)+20+e [−5.12, 5.12]
f11(x)=14000i=1Nxi2i=1Ncos(xii)+1 [−32, 32]
f12(x)=πn{10sin(πy1)+i=1n1(yi1)2[1+10sin2(πyi+1)]+(yn1)2}+i=1nu(xi,10,100,4) [−600, 600]
u(xi,a,k,m)={k(xia)m,xi>a0,a<xi<ak(xia)m,xi<ayi=1+xi+14 [−50, 50]
f13(x)=0.1{sin2(3πx1)+i=1n(x11)2[1+sin2(3πxi+1)]+(xn1)2[1+sin2(2πxn)]}+i=1nu(x1,5,100,4) [−50, 50]
Table 1

The benchmark functions

The proposed algorithm was run 30 times on each benchmark functions. The average results are summarized in Table 2. The performance of the proposed algorithm is compared with two EAs: DE and fast evolutionary programming (FEP) and compared with swarm-based algorithm: PSO. The results of DE, FEP and PSO were taken from the results reported in Saremi et al. [12].

Functions DE/GOA DE FEP PSO
f1 2.35E–220 8.20E–14 5.70E–04 1.36E–04
f2 4.28E–11 1.50E–09 8.10E–03 4.21E–02
f3 3.50E–226 6.80E–11 1.60E–02 70.12562
f4 0 0 3.00E–01 1.086481
f5 28.98017 0 5.06 96.71832
f6 0 0 0 1.02E–04
f7 2.11E–05 4.63E–03 1.42E–01 1.23E–01
f8 −8831.53 −11080.1 −12554.5 −4841.29
f9 6.76E–10 69.20 4.60E–02 46.70423
f10 4.44E–16 9.70E–08 1.80E–02 2.76E–02
f11 0 0 1.60E–02 9.22E–03
f12 2.87E–09 7.90E–15 9.20E–06 6.92E–03
f13 3.01E–08 5.10E–14 1.60E–04 6.68E–03
Table 2

The experimental results

According to the results of Table 2, DE/GOA is able to provide very competitive results. This algorithm outperforms all others in f1, f2, f3, and f7. Both DE/GOA and DE were successful to find the optimal solution in f4 and f6. In the function f5, only DE could solve the optimum. Therefore, these results show the performance of DE/GOA in terms of exploiting the optimum.

The multimodal functions have many local optima with the number increasing exponentially with dimension. This makes them suitable for benchmarking the exploration ability of an algorithm. The experimental showed that DE/GOA is able to provide very competitive results on the multimodal benchmark functions as well.

The experimental results show that the global search algorithms which added grasshopper optimization algorithm could increase convergence rate because the characteristic of GOA is both exploration and exploitation. Meanwhile the results of multimodal functions are not perfect that may be due to the multimodal functions have many increasing number of local optimum.

5. CONCLUSION

This work proposed to improve the DE algorithm performance with integrated the GOA. Thirteen benchmark functions were employed the performance in terms of exploitation and exploration. The results showed that DE/GOA was able to provide competitive results compared to well known heuristics such as DE, FEP and PSO.

Authors Introduction

Dr. Duangjai Jitkongchuen

She graduated Ph.D. in Information Technology from King Mongkut’s Institute of Technology Ladkrabang (KMITL) in 2014 and now she is working as Assistant Professor at Dhurakij Pundit University, College of Innovative Technology and Engineering as a deputy dean administration. Her research area is about machine learning, data mining and bio-inspired algorithms.

Ms. Udomlux Ampant

She graduated Master of Business Administration (Marketing) Dhurakij Pundit University. She is an lecturer in the College of Innovative Technology and Engineering in Dhurakij Pundit University.

REFERENCES

[1]JH Holland, Genetic algorithms, Sci. Am., Vol. 267, 1992, pp. 66-72.
[6]FF Moghaddam, RF Moghaddam, and M Cheriet, Curved space optimization: a random search based on general relativity theory, 2012. arXiv:1208.2214 (https://arxiv.org/abs/1208.2214).
Journal
Journal of Robotics, Networking and Artificial Life
Volume-Issue
5 - 3
Pages
165 - 168
Publication Date
2018/12/01
ISSN (Online)
2352-6386
ISSN (Print)
2405-9021
DOI
10.2991/jrnal.2018.5.3.5How to use a DOI?
Copyright
© 2018 The Authors. Published by Atlantis Press SARL.
Open Access
This is an open access article under the CC BY-NC license (http://creativecommons.org/licences/by-nc/4.0/).

Cite this article

TY  - JOUR
AU  - Duangjai Jitkongchuen
AU  - Udomlux Ampant
PY  - 2018
DA  - 2018/12/01
TI  - Integrated Optimization of Differential Evolution with Grasshopper Optimization Algorithm
JO  - Journal of Robotics, Networking and Artificial Life
SP  - 165
EP  - 168
VL  - 5
IS  - 3
SN  - 2352-6386
UR  - https://doi.org/10.2991/jrnal.2018.5.3.5
DO  - 10.2991/jrnal.2018.5.3.5
ID  - Jitkongchuen2018
ER  -