Integrated Optimization of Differential Evolution with Grasshopper Optimization Algorithm
- DOI
- 10.2991/jrnal.2018.5.3.5How to use a DOI?
- Keywords
- Meta-heuristic; differential evolution algorithm; grasshopper optimization algorithm; optimization
- Abstract
This paper proposes a scheme to improve the differential evolution (DE) algorithm performance with integrated the grasshopper optimization algorithm (GOA). The grasshopper optimization algorithm mimics the behavior of grasshopper. The characteristic of grasshoppers is slow movement in the larval stage but sudden movement in the adulthood which seem as exploration and exploitation. The grasshopper optimization algorithm concept is added to DE to guide the search process for potential solutions. The efficiency of the DE/GOA is validated by testing on unimodal and multimodal benchmarks optimization problems. The results prove that the DE/GOA algorithm is competitive compared to the other meta-heuristic algorithms.
- Copyright
- © 2018 The Authors. Published by Atlantis Press SARL.
- Open Access
- This is an open access article under the CC BY-NC license (http://creativecommons.org/licences/by-nc/4.0/).
1. INTRODUCTION
Meta-heuristic algorithm techniques constitute range from simple local search procedures to complex learning processes. The basic concepts of meta-heuristics permit an abstract level description. Nature-inspired meta-heuristic can be grouped in three main classes: Evolution-based, physics-based and swarm-based.
Evolution-based algorithms are inspired by the concepts of nature evolution. The population in search process are randomly generated which are developed over subsequent generation. The next generation is created by combination of the individuals in the previous generation. These provide the population to be optimized over the course of generations. The most popular evolution-based algorithm is genetic algorithms (GA) [1]. GA has a process of fitness-based selection and recombination to produce a successor population for the next generation. During recombination, child chromosomes are produced by recombined genetic material from parent chromosomes are selected. As this process is repeated, a sequence of consecutive generations evolves and the average fitness of the chromosomes inclines to increase until some stopping criterion is reached. Other popular algorithms such as differential evolution (DE) algorithms [2], the principal difference between genetic algorithms and differential evolution is that DE use mutation as the primary search mechanism, while GA use crossover as the probabilistic mechanism to exchange of information among solutions to locate better solutions. Biogeography-based optimizer (BBO) [3] algorithm is applied from nature phenomenon regarding the distribution of living creatures in various islands. In this field of study different ecosystems are explored for finding the relations between different species in terms of immigration, emigration, and mutation. The main inspiration of the BBO algorithm is to reach a stable situation based on migration and mutation in the evolution of ecosystems.
The second class of meta-heuristic is physics-based algorithms that imitate the physical rules. For example, gravitational search algorithm (GSA) [4], GSA is a nature-inspired conceptual framework based on gravitational kinematics, a branch of physics that models the motion of masses moving under the influence of gravity. Galaxy-based search algorithm [5] is inspired on the spiral arm of spiral galaxies to search its surrounding. This spiral movement is improved by chaos to escape from local optimums. Curved space optimization [6] inspired by general relativity theory is used to enhance the efficiency of a simple random search, and convert it to a very robust optimization tool.
The third class of meta-heuristic is swarm-based algorithms that imitate the social behavior of swarms, herds, flocks, or schools of animals. The most popular algorithms are particle swarm optimization (PSO) [7]. PSO inspired by bird flocking or fish schooling. Each particle will be iteratively updated to improve its solution during movement around the search space. The result leads to the best value of global solution. Other popular swarm-based algorithms are bat algorithms (BA) [8]. BA is inspired by echolocation behavior of microbats to avoid obstacles, find their home, and detect their prey in the darkness. Firefly algorithm (FA) [9] applied from the flashing light of fireflies when they want to mate or to warn others about predators.
The new meta-heuristic has been focused on this paper is differential evolution algorithm which is a global search optimization algorithm and improved performance of it by added grasshopper optimization algorithm (GOA) to increase convergence rate.
The rest of the paper is structured as follows. Section 2 describes some backgrounds of the differential evolution algorithm. In Section 3, the proposed algorithm is presented. Section 4 shows the experimental results and the conclusions will be discussed in Section 5.
2. THE DIFFERENTIAL EVOLUTION ALGORITHM
The differential evolution algorithm optimizes a problem by sustaining a population of candidate solutions and creating new candidate solutions by combining existing ones according to its simple procedure, and then keeping whichever candidate solution has the best score or fitness on the optimization problem at hand.
The procedure of differential evolution algorithm starts to initialize target vectors
The target vectors are used to generate the mutant vectors
The last step is the selection operation to choose a better vector for next generation. The new generations are selected by comparing the fitness value between the target vector and the trial vector. The selection is shown in Eq. (3).
3. THE PROPOSED ALGORITHM
The schema of DE/current-to-best/1 in mutation operation is updated by splitting F mutation parameter to λ and F. The new mutation schema is shown in Eq. (4).
After that, the crossover operation is rearranged. The vectors that created from the target vectors and the mutant vector is called the prime vectors (U′i). The prime vectors is used to generate the trial vectors by comparing the fitness value between the prime vectors and the mutant vectors, otherwise the trial vectors is generated by grasshopper optimization algorithm.
The grasshopper optimization algorithm was proposed in Saremi et al. [10] mimics the behavior of grasshopper. The characteristic of the swarm in the larval stage is slow movement and small steps but is long range and suddenly movement in adulthood. These two characters seem as exploration and exploitation. The search agents are encouraged to move suddenly in exploration and conduce to move locally during exploitation. These two functions as well as target seeking are performed by grasshopper naturally.
The proposed algorithm takes the feature of movement of the grasshopper to generate a new value for the trial vectors is showed in Eq. (8).
4. THE EXPERIMENTAL RESULTS
The proposed algorithm has been evaluated performance with thirteen benchmark functions [11]. The test functions are unimodal (f1 − f7) and multimodal (f8 − f13). The details of benchmark functions are shown in Table 1. The mutation parameters set λ, F = 0.5, the crossover parameter set CR = 0.9 and population size set NP = 100.
Functions | Range |
---|---|
[−100, 100] | |
[−100, 100] | |
[−10, 10] | |
f4(x) = maxi {|xi|, 1 ≤ i≤ N} | [−100, 100] |
[−100, 100] | |
[−30, 30] | |
[−100, 100] | |
[−1.28, 1.28] | |
[−500, 500] | |
[−5.12, 5.12] | |
[−32, 32] | |
[−600, 600] | |
[−50, 50] | |
[−50, 50] |
The benchmark functions
The proposed algorithm was run 30 times on each benchmark functions. The average results are summarized in Table 2. The performance of the proposed algorithm is compared with two EAs: DE and fast evolutionary programming (FEP) and compared with swarm-based algorithm: PSO. The results of DE, FEP and PSO were taken from the results reported in Saremi et al. [12].
Functions | DE/GOA | DE | FEP | PSO |
---|---|---|---|---|
f1 | 2.35E–220 | 8.20E–14 | 5.70E–04 | 1.36E–04 |
f2 | 4.28E–11 | 1.50E–09 | 8.10E–03 | 4.21E–02 |
f3 | 3.50E–226 | 6.80E–11 | 1.60E–02 | 70.12562 |
f4 | 0 | 0 | 3.00E–01 | 1.086481 |
f5 | 28.98017 | 0 | 5.06 | 96.71832 |
f6 | 0 | 0 | 0 | 1.02E–04 |
f7 | 2.11E–05 | 4.63E–03 | 1.42E–01 | 1.23E–01 |
f8 | −8831.53 | −11080.1 | −12554.5 | −4841.29 |
f9 | 6.76E–10 | 69.20 | 4.60E–02 | 46.70423 |
f10 | 4.44E–16 | 9.70E–08 | 1.80E–02 | 2.76E–02 |
f11 | 0 | 0 | 1.60E–02 | 9.22E–03 |
f12 | 2.87E–09 | 7.90E–15 | 9.20E–06 | 6.92E–03 |
f13 | 3.01E–08 | 5.10E–14 | 1.60E–04 | 6.68E–03 |
The experimental results
According to the results of Table 2, DE/GOA is able to provide very competitive results. This algorithm outperforms all others in f1, f2, f3, and f7. Both DE/GOA and DE were successful to find the optimal solution in f4 and f6. In the function f5, only DE could solve the optimum. Therefore, these results show the performance of DE/GOA in terms of exploiting the optimum.
The multimodal functions have many local optima with the number increasing exponentially with dimension. This makes them suitable for benchmarking the exploration ability of an algorithm. The experimental showed that DE/GOA is able to provide very competitive results on the multimodal benchmark functions as well.
The experimental results show that the global search algorithms which added grasshopper optimization algorithm could increase convergence rate because the characteristic of GOA is both exploration and exploitation. Meanwhile the results of multimodal functions are not perfect that may be due to the multimodal functions have many increasing number of local optimum.
5. CONCLUSION
This work proposed to improve the DE algorithm performance with integrated the GOA. Thirteen benchmark functions were employed the performance in terms of exploitation and exploration. The results showed that DE/GOA was able to provide competitive results compared to well known heuristics such as DE, FEP and PSO.
Authors Introduction
Dr. Duangjai Jitkongchuen
She graduated Ph.D. in Information Technology from King Mongkut’s Institute of Technology Ladkrabang (KMITL) in 2014 and now she is working as Assistant Professor at Dhurakij Pundit University, College of Innovative Technology and Engineering as a deputy dean administration. Her research area is about machine learning, data mining and bio-inspired algorithms.
Ms. Udomlux Ampant
She graduated Master of Business Administration (Marketing) Dhurakij Pundit University. She is an lecturer in the College of Innovative Technology and Engineering in Dhurakij Pundit University.
REFERENCES
Cite this article
TY - JOUR AU - Duangjai Jitkongchuen AU - Udomlux Ampant PY - 2018 DA - 2018/12/01 TI - Integrated Optimization of Differential Evolution with Grasshopper Optimization Algorithm JO - Journal of Robotics, Networking and Artificial Life SP - 165 EP - 168 VL - 5 IS - 3 SN - 2352-6386 UR - https://doi.org/10.2991/jrnal.2018.5.3.5 DO - 10.2991/jrnal.2018.5.3.5 ID - Jitkongchuen2018 ER -