Next Article in Journal
Li2CO3 as Protection for a High-Temperature Thermoelectric Generator: Thermal Stability and Corrosion Analysis
Previous Article in Journal
Bridging the Gaps in Traceability Systems for Fresh Produce Supply Chains: Overview and Development of an Integrated IoT-Based System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparative Analysis of Low Discrepancy Sequence-Based Initialization Approaches Using Population-Based Algorithms for Solving the Global Optimization Problems

by
Waqas Haider Bangyal
1,
Kashif Nisar
2,*,
Ag. Asri Bin Ag. Ibrahim
2,
Muhammad Reazul Haque
3,
Joel J. P. C. Rodrigues
4,5 and
Danda B. Rawat
6
1
Department of Computer Science, University of Gujrat, Gujrat 50700, Pakistan
2
Faculty of Computing and Informatics, Universiti Malaysia Sabah, Jalan UMS, Kota Kinabalu 88400, Malaysia
3
Faculty of Computing & Informatics, Multimedia University, Cyberjaya 63100, Malaysia
4
Post-Graduation Program on Electrical Engineering, Federal University of Piauí (UFPI), Teresina 64049-550, PI, Brazil
5
Instituto de Telecomunicações, 6201-001 Covilhã, Portugal
6
Department of Electrical Engineering and Computer Science, Howard University, Washington, DC 20059, USA
*
Author to whom correspondence should be addressed.
Submission received: 16 April 2021 / Revised: 10 May 2021 / Accepted: 13 May 2021 / Published: 18 August 2021
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
Metaheuristic algorithms have been widely used to solve diverse kinds of optimization problems. For an optimization problem, population initialization plays a significant role in metaheuristic algorithms. These algorithms can influence the convergence to find an efficient optimal solution. Mainly, for recognizing the importance of diversity, several researchers have worked on the performance for the improvement of metaheuristic algorithms. Population initialization is a vital factor in metaheuristic algorithms such as PSO and DE. Instead of applying the random distribution for the initialization of the population, quasirandom sequences are more useful for the improvement the diversity and convergence factors. This study presents three new low-discrepancy sequences named WELL sequence, Knuth sequence, and Torus sequence to initialize the population in the search space. This paper also gives a comprehensive survey of the various PSO and DE initialization approaches based on the family of quasirandom sequences such as Sobol sequence, Halton sequence, and uniform random distribution. The proposed methods for PSO (TO-PSO, KN-PSO, and WE-PSO) and DE (DE-TO, DE-WE, and DE-KN) have been examined for well-known benchmark test problems and training of the artificial neural network. The finding of our techniques shows promising performance using the family of low-discrepancy sequences over uniform random numbers. For a fair comparison, the approaches using low-discrepancy sequences for PSO and DE are compared with the other family of low-discrepancy sequences and uniform random number and depict the superior results. The experimental results show that the low-discrepancy sequences-based initialization performed exceptionally better than a uniform random number. Moreover, the outcome of our work presents a foresight on how the proposed technique profoundly impacts convergence and diversity. It is anticipated that this low-discrepancy sequence comparative simulation survey would be helpful for studying the metaheuristic algorithm in detail for the researcher.

1. Introduction

The Meta-heuristic algorithms are an integral part of modern optimization. In the last couple of decades, a wide range of metaheuristic algorithms have arisen, and many meta-heuristics are incredibly popular, such as particle swarm optimization. Initialization of the population plays a vital role in Meta-heuristic algorithms. Generally, a Population-based algorithm is used for global optimization problems [1]. Meta-heuristics are problem-saving generalizations, typically stochastic as well as the most common method of minimizing global optimization. In the latest times’ researchers have shown an increasing curiosity in population-based stochastic exploration strategies to tackle the problems of global optimization. Meta-heuristic can be used to optimize combinations in an unconfined search area in which an optimum solution is sought. Another example of the problem is the TSP, where a candidate’s quest for a solution increases exponentially because the problem increases, and more solutions are not possible [2]. Optimization is an attempt of acquiring the optimum solution of a problem under given instances. All systems which might be optimized have an objective function and many decision variables that affect the functions. The important undertaking challenges of optimization is to reduce wastage time or exploit the preferred advantage of a given engineering system [3]. A preference to utilize existing sources for increasing the optimization strategies within the exceptional feasible way. Optimization techniques may be described as a procedure that accomplished superior solutions which fulfill the given objective feature [3]. Optimization refers to the observation of those problems wherein one seeks to reduce or maximize a characteristic through technically choosing the variables’ values within their allowed collections [4]. All possible benefits are recognized as acceptable solutions, while extraordinary output is considered to be the best solution. In recent few years, to tackle optimization complications, scientific society is aware of the development of techniques influenced by nature. The growing intelligence level of groups of basic independent agents is swarmed intelligence. 2 Swarm Intelligence (SI) is a field of Artificial Intelligence (AI), introduced by Gerardo Beni and Jing Wang. It is inspired by the nature and collective behavior of all individuals [5]. SI has no integrated but self-organized control structure. In which all Individuals follow certain agents that closer to the optimal outcome. An independent agent is a control system that communicates along with its ecosystem, obviously comprised of other agents, but behaving comparatively individually of all the other agents. The independent agent does not follow other participant’s instructions or a global plan. In nature, such systems are generally used to solve problems such as effective searching for food, target escaping, or colonies relocation. Generally, the information is stowed by partner agents that are enclosed in the environment, such as through the use of pheromones in flies and ants and closeness in the birds and fish [6]. Here we take the birds’ example that flies in a decentralized control structure, in which all the birds follow those birds that are nearer to the food. The SI library can be used for the acquisition of transit material, communication, medical database classification, dynamic control, heating plans, tracking problems, and predictions that have been valuable and intriguing, just as somewhat compromising, although SI works for the new innovative advancement [7]. Several influential population-based metaheuristics techniques had been produced and used successfully to solve many complex optimization problems. Some majors are Particle-Swarm-Optimization (PSO) [8], Bat algorithm [9], Ant Colony Optimization (ACO) [10], Bee Colony Algorithm (BCA) [11], etc. The focus of this research on Meta-heuristic global optimization with the Particle Swarm Optimization (PSO) technique in the literature of Swarm Intelligence originally by Kennedy and Eberhart [8]. Generally, PSO can be extended like many evolutionary computational methodologies to resolve most scalability issues and problems which can be converted into optimization problems [12]. Original meaning this algorithm is based on the flocking of birds and fishing school movement behavior while searching for food they scattered and place on the different locations where they find the food. While searching from one place to another, if any agent smells the food, they transmit information to each other, all the birds follow him to get the food without wasting time to reach the place where the food is. The most optimum solution of PSO can be worked out by the collective behavior of all individuals [13]. In this thesis work, the focus of study on Particle-Swarm-Optimization (PSO) that was initially introduced by James Kennedy and Russel Eberhart is a meta-heuristic global optimization strategy. Generally, PSO can be extended like many evolutionary computational methodologies to resolve most scalability issues and complex problems which altered into optimization problems. This algorithm’s significance is based on the flocking of the birds and the movement of fishing schools in their search for food that anywhere found [12]. When looking from one place to the next, if any agent detects the food, without wasting any time to fetch the food, all pursue the agent to reach the place. The collective behavior of all individuals can be the optimum PSO solution The PSO algorithm firstly searches for the optimum value of each particle and finds according to the minimum range values with the corresponding support and confidence, later the data is converted to binary values. Areas where other PSOs have shown specific consent, include multi-modal issues and concerns in which no advanced technique is accessible or for which unsatisfying outcomes are offered by all specialized techniques. Figure 1 illustrates the evolution of the particles in such a simplistic, 1-dimensional search area. Since 1995 its introduction, Particle-swarm-optimization (PSO) has several improvements. Researchers learned about this method, many new systems also have been inferred and new applications have been developed. Optimization of particle swarm has been used in a variety of applications [12], the most latent are system design applications, clustering, multi-objective optimization problems solving applications, pattern recognition application, power generation applications, sensor networks and neural network, fuzzy and neuro-fuzzy structures and controller, music generation applications, prediction and forecasting, detection of faults and recovery applications, games applications, optimization of communication networks, biological robotic applications, design/ optimization of electrical motors and engines, simulation and identification applications, computer graphical visualization applications, design and reformation of electricity networks and load dispatching application, electronics and electromagnetics applications, finance and economics, images and videos analysis applications, security and military applications, medicinal and pharmacological applications, data mining, signal processing, and decision making applications [12].
Above we have discussed some applications where it can be used. Till from its birth, it has lots of improvement and also integrated with another Mata-heuristic algorithm to improve its performance. Due to the easiness of implementation and the ability to promising optimization, PSO is successfully implemented in many fields [14]. However, the PSO has faced some problems such as the high complexity of the computation, premature convergence, and diversity of the algorithm. In some cases, the entire search area cannot explore a small population size with large iteration but can exploit some small area of search space by doing a local search, so a premature convergence problem happens [15]. Particularly to solving these complex problems, many researchers are devoted to dealing with this issue. The main reason for the premature convergence issue is the lack of diversity in the population [16]. It is very important to make the diversity of the population perfect.
Above we have discussed all the applications or fields where it can be used. PSO classifier is not only a tool for solving optimization problems but also a tool to comprise individuals as well as artificial agents socio-cognition related to social psychology principles. A PSO system integrates local search techniques with global search techniques to seek a sort of balance between exploration and exploitation [17]. The main concept of the exploration and exploitation are searching for the unknown solution points are discuss given below. Exploration means explore the whole search space to search for a new individual that far from the current individual. It has less chance to trap in the given search space. Exploitation is fewer exploration means explore the limited or restricted area. It utilizing the solution points in hand means to search the new individual in the search space that is nearby the current individual [18]. Premature convergence means that a group of the population while moving in the search space to find optimal solution converged too early [19].
PSO algorithm uses real number coding. Since there’s no selection, crossover, and mutation, the algorithm’s frame is relatively easy and the velocity of deployment is very quick. However, if some particle in the swarm finds the current optimal position, the other particles will quickly try to gather near it. If the position is a maximum of local, then the swarm of particles will not be able to find it again. Therefore, the algorithm is drowned in the maximum of the local and premature convergence occur [20]. From different perspectives, in each iteration, members of the population-based search algorithm understand the conceptions of exploring and exploiting in three phases: self-creation, collaboration, and competency. In the self-creation step, each individual improves their performance. In the collaboration step, each individual interacts with the other by transferring the information. Finally, with incompetency, each strives to suffer [21]. Generally, these phases are stochastic forms and can be utilized in different ways. These are the core concepts beyond nature-inspired population-based heuristic algorithms. These ideas help an algorithm find the best overall solution. As Swarm Intelligence (SI) methodology, due to the easiness of implementation and the ability to promising optimization, PSO is successfully implemented in a comprehensive scientific analysis and engineering solution.
However, the PSO has faced some problems such as the high complexity of the computation, premature convergence, and diversity of the algorithm. In certain cases, a limited population size combined with a large number of iterations is unable to cover the entire search space but can exploit some small area of search space by doing a local search, so a premature convergence problem happens. Particularly to solving these complex problems, many researchers are devoted to dealing with this issue [13]. Like other algorithms of empirical evolution, during the evolution process, the main cause for the premature convergence of Basic PSO is the rapid reduction of diversity of the population. It is very important to make the diversity of the population perfect. In general, there are some techniques for maintaining diversity. When an algorithm is used to measure the diversity of the population, it is a consistent measure to compare the diversity estimated value to different degrees. But the limit value constantly affects the properties of the PSO algorithm [22]. If the current zone is a local solution, the low diversity will obtain total collapse and finally, inflammation detect [23]. Consequently, many researchers are committed to increasing diversity to improve the quality of PSO solutions, especially for multi modal and multipurpose functions.
Another issue that PSO facing is not finding the optimum solution. While dealing with complex problems, swarms get stuck in local optima. The aspect of those new ideas include improvement in searching manners, increase the diversity in the population, modified existing strategies, and introduced new ideas. Therefore, we will introduce the new version of the PSO algorithm for finding global minima. This PSO can easily explore the whole search area using an effective diversity enhancement mechanism [9]. In this thesis, it is proposed to use the modified PSO strategy with a novel initialization approach to solving complex problems. In general, random starts are used to produce the initial population when. Usually, random initialization is used to produce an initial population while priority information is missing. The word random" must no longer obscure the readers within the QRS term. Such is neither pseudorandom nor actual random sequences. Certainly, QRSs are completely deterministic and in algorithms no random element is concerned but optimization problems are very useful [24]. The most commons are Sobol, Vander Corput, WELL, Torus, Halton, and Faure. PRNGs produce a series of numbers that closely resemble random numbers. these random numbers are efficient, deterministic, and periodic. The most commons are LCG, MCG, LFG, and ICG. These sequences were introduced to initialize the swarm as well as the quantum results indicate a significant development concerning traditional BPSO, using uniform random numbers distributed [25]. QRS and PRNGs enhanced population performance by initializing the QRS population.
The rest of the paper is structured as follows: Related work is presented in Section 2. Methodology is elaborated in the Section 3. Experimental setup is described in Section 4. Simulation results and discussions are presented in Section 5. Comparison of PSO and DE regarding data classification presented in Section 6, and lastly, the conclusion and future work described in the Section 7.

2. Related Work

The idea of using a random number generator for initializing the population into a multidimensional search space is not new. An essential step in any metaheuristic algorithm is to initialize its population correctly. If the initialization is not proper, then it may go on to search in unnecessary areas and may fail to explore the optimum solution. Proper initialization is very important for the metaheuristic algorithm for its performance. Many research studies proposed different variants based on initialization techniques.
Initialization of the swarm in a good way helps the PSO to search more efficiently [26]. In this work, the initialization of swarm with nonlinear simplex method (NSM) has been conducted. NSM requires only function evaluations without any derivatives for computation. NSM starts with initial simplex and produces sequence of steps moving the highest function value vertex in the opposite direction of the lowest one. They initialized the particle with the initial simplex in the D-dimensional search areas D+1, vertices of the simplex are D+1 particle of the swarm, and applied the MSN method for N-D+1 steps for N size swarm. In this way, each particle in the swarm has the information of the region. Lastly, they compared their results with simple PSO and found significant improvement.
This variant was introduced by Mark Richards and Dan Ventura in 2004. In their work [27], they proposed to use centroidal Voronoi tessellations for initializing the swarm. Voronoi tessellations is a technique of partitioning any region into compartments; each partition contains a group of generators. Each partition is associated with one generator, and it consists of all the particles closer to that generator. In the same way, the generators are selected for the initial position of the particle. In this way, they initialized the particle swarm optimization algorithm. They compared it with basic SPO on many benchmark functions and found improved performance in high-dimensional spaces.
Halton sampling was introduced by Nguyen Xuan Hoai, Nguyen Quang Uy, and R.I. McKay in 2007 [28,29]. Halton sequence is a low-discrepancy deterministic sequence used to generate a point in space. Halton sequence is not entirely random. To randomize X, Wang and F. J. Hickernell proposed a new function called randomize Halton sequence by using von Neumann–Kakutani transformation. They used this sequence to initialize the global best of the PSO. They performed a test on various benchmark functions and compared the result with the PSO and initialized with uniform random numbers. They found better performance, especially for complex and smaller populations.
VC-PSO was introduced by Millie Pant et al. in 2008. They used Van der Corput sequence for initializing the swarm for large dimensions’ search [30,31]. The Van der Corput and Sobol sequence generate a semirandom number, which is more suitable for computational purposes. They tested the new variant with different benchmark functions and compared the result with BPSO and SO-PSO and found significant improvement, especially for large search space dimensions. The main purpose of this variant is to see the performance of large search space problems. That is why they used search space with different sizes ranging [−5.12, 5.12] to [−1000, 1000]. The performance of the algorithm is showing prominence when the dimension increases to [−100, 100].
SMPO was introduced by Millie Pant et al. in 2008. They used a quasirandom Sobol sequence to initialize the particles instead of standard random numbers [25,32]. They used a new operator called a systematic mutation operator, which is used to improve the performance of the PSO. The new operator, instead of using the standard random number, use a quasirandom Sobol sequence to initialize the swarm as the QRS is less random as compared to pseudorandom sequences, which is helpful for computational methods. They proposed two variants SM-PSO1 and SM-PSO2. The main difference between the two versions is that in MSPSO1, the best particle is mutated, while in MS-PSO2, the worst particle is mutated. They found better results compared to BPSO and other variants.
This work is done by [33]. In this paper, researchers have proposed a new method of initialization. In their work, they have added the functionality to automatically detect when a particle is prematurely converged and initializes the swarm. They also added features to redesign the inertia weight to balance the searching ability globally and locally. They named it as IAWPSO.
This variant is proposed by P. Murugan in 2012 and applied on the transmission expansion problem to decide the installation of new circuits in increased usage of electricity and found this variant fruitful [34]. In this work, he used a new initialization technique called population monitored for complementary magnitudes initialization. In initialization, he used decision variables. All particles are initialized with an integer within the limit of the upper and lower values of the decision variable in such a way that each particle should be unique. The initial population is created in such a way that each particle can have the ability of the possible solution, and they are unique. Almost 50% of the particles are opposite to another 50%, considering the lower and upper limit of the decision variable. The important thing in this initialization is to maintain uniqueness and diversity among the particles of the swarm generated initially.
SIS_PSO was introduced by [35]. In this paper, the authors introduced a new initialization technique named space-based initialization strategy. In this work, they broke down each dimension of the search area into two segments: S1i and S2i. The border of the areas is [li, (li + ui)/2] and [(li + ui)/2, ui], and each segment is linked with a probability and initialized with 0.5. They applied SIS-PSO on 13 functions and compared the result with GPSO and CLPSO and found significant improvement.
Jensen et al. [36] describes that problem of finding the global minima or maxima in single real variable f(x) of continuous function is an equally important research domain for science and engineering. A number of methods have been proposed so far to properly solve this problem of finding global minima or maxima. All these methods have been generally categized into two sets or groups as deterministic or stochastic. In this paper, the author catered to these problems by introducing a novel approach as mix of both deterministic and stochastic approaches, by finding the precision of the deterministic approaches along with the localization ability of stochastic approaches. The proposed novel approach is aligned with genetic algorithm (GA).
This variant was introduced by [37]. In this work, they introduced a new variant of PSO called polar PSO. They explained that mostly the distortion occurred due to the polar particles. Hence, they introduced a new method for the reinitialization of the polar particles by redefining the distance based on the dimensionality of the point. By using this method, they removed the distortion occurring during the computation. They compared the results with BPSO and found some improvement.
This variant was proposed by [38]. In this work, they used the Nawaz–Enscore–Ham heuristic technique to initialize the swarm. This variant is named PHPSO. The sequence generated by NEH jobs is placed in ascending order of the sums of their total flow time. To construct a job sequence, it depends on its initial order. The minimum TFT sequence is the current sequence for the upcoming iteration among all the sequences. The resulting population generated by the NEH method is used to initialize the population of PSO. They applied this algorithm for the no-wait flow shop scheduling problem. They compared the result with DPSO and HPSO and found a comparatively better result.
This work is done by [39]. They proposed a new algorithm combining Adoptive PSO with backpropagation (BP) to train weights of neural network. In this work, they introduced different selection techniques to select inertia weight. They named it as PSO-BP. In the end, the BP algorithm was used to search around the global optimum. They compared the result with adaptive particle swarm optimization (APSOA) and back propagation algorithm and found better results.
This work is done by [40]. In this work, they used PSO to optimize weight and architecture of MLP feed-forward neural network. They have used PSO algorithms twice: the inner one and the outer one. The inner one was used to optimize the weight, and the outer one was used for architecture optimization. In the proposed technique, the outer PSO searches the hidden nodes of each hidden layers of the MLP networks. To improve the generalization, they used weight decay heuristics in the inner PSO. The performance of the proposed technique was applied to the famous benchmark classification problem in the medical field. The result shows better generalization for all data sets.
PSO-ANN is proposed by [41]. In this work, the authors showed a better performance of ANN by using PSO to adjust the weights. In the proposed technique, both the velocity and position of each particle was initialized randomly in a specific dimension to train the PSO-ANN based on particle position and calculated its fitness. They saved the current position and fitness as history. The individual best among all the particle was considered to be the best. The new set of position generated by this method was used for learning error generation. Every particle’s position and velocity was updated according to its personal best and global best. This helped the biases and weights trapping in a local minimum. The performance of PSO-ANN is compared with many backpropagation algorithms such as Levenberg–Marquardt and gradient descent algorithms resulting in better performance.
A new variant of PSO combining with stochastic gradient decent is proposed by [42] and named the PSO-SGD algorithm for training convolution neural network [42]. The proposed technique worked into two phases. PSO was used to train and initialize the CNN parameters in the first phase. When it showed, slow progress of PSO for a few iteration SGD was used in the second phase. In addition, they used PSO combining with genetic algorithm (GA), which helped the particle for simulation and overcame the slowness of SGD. They applied the new algorithm on a different benchmark data set and performed well for three different data sets. The proposed technique avoided the occurrence of local optimum and premature saturation, as it was in the known problem by using any single algorithm.
The authors in [43] examined the impact of initiating the initial population by excluding traditional techniques such as random numbers. The authors applied the nonlinear simplex method for generating the initial population of DE, where the proposed algorithm was termed NSD. The working of the proposed algorithm is measured with twenty benchmark functions and compared with standard DE and opposition-based DE (ODE) algorithm. Numerical results illustrate that the proposed technique enhances the convergence rate.
During the image segmentation process, to solve the issue of thresholding, an enhanced variant of standard DE algorithm with a local search (termed as LED) and low-discrepancy sequences is introduced [44]. Experimental results conclude that the performance of the introduced algorithm is superior for finding the optimum threshold.
For the steelmaking continuous (SCC) problem, in [45], the authors presented a novel enhanced technique of DE based on the two-step procedure for producing an initial population, as well as, a novel mutation approach. Furthermore, an incremental methodology for generating the initial population is also incorporated in DE to handle dynamic events. Computational experiments conducted with the presented approach show the effectiveness of the presented approach than others.
In the time-series prediction, the backpropagation neural network (BPNN) gets stuck into local optima. In [46], a hybrid technique having adaptive DE (ADE) and BPNN, termed ADE-BPNN, is developed to enhance the accuracy of BPNN. To validate the effectiveness of the proposed technique, two real-world time-series data sets are used. The results show that the proposed approach gives higher efficiency.
The authors in [47] proposed an improved differential evolution BA algorithm by combining BA algorithm with DE for optimizing NN for optimizing fuzzy NN to predict the traffic of the network. Due to the problems of DE such as the movement of premature convergence and inactive convergence rate, Gaussian disturbance crossover operator and adaptive mutation operator are used in an introduced variant of DE. The improved operators are utilized to design the crossover parameter and enhance the mutation of traditional DE. To test the performance of the proposed modification, four traditional functions and network traffic used. The results describe that the proposed variant of DE improves the prediction accuracy and the convergence rate as compared to other techniques.
For classification, a modified DE-trained Pi-sigma network (PSN) is presented [48]. The presented technique is an enhanced version of a novel mutation and standard DE/rand/1/bin algorithm and, in addition to this, crossover approaches utilized with regards to both exploitation and exploration. The presented approach for pattern classification is validated through three famous real-world classification problems taken from the UCI repository. The experimental results of the presented method is verified with two well-known algorithms: chemical reaction optimization and DE/rand/1/bin algorithms. The results depict that the presented approach is superior.

3. Methodology

The most important step in any evolutionary computing algorithm is to initialize its population properly. If the initialization is not proper, then it may go to search in unnecessary areas and may fail to search the optimum solution. Proper initialization is very important for any metaheuristic algorithm for its performance. A metaheuristic is random; therefore, it does not have a specific pattern to ensure the optimum global point. Therefore, by taking advantage of this randomness and considering this fact, we have proposed three novel quasirandom initialization strategies called WELL sequence, Knuth sequence, and Torus sequence to initialize the population in the search space. We initialized PSO and DE algorithm with these proposed pseudorandom strategies (WELL sequence, Knuth sequence, and Torus sequence). We have compared the novel techniques with the simple random distribution and family of low-discrepancy sequences on several unimodal and multimodal complex benchmark functions and training of the artificial neural network. A brief description of quasisequence approaches and proposed algorithms using WELL sequence, Knuth sequence, and Torus sequence for PSO, and DE is discussed below.

3.1. Random Number Generator

As discussed above, inbuilt library function is used, R a n d ( x m i n , x m a x ) to generate the mesh of numbers randomly at uniform locations [49]. The probability density function of a continuous uniform distribution identifies the impact of uniformity on any sequence. The probability density function can be characterized as given:
f ( t ) = { 1 p q   f o r   p < t < q 0           f o r   t < p   o r   t > q
where p and q represent the parameter of maximum likelihood. The value of f ( t ) is useless at the boundary of p and q due to the 0 effect on the integrals of f ( t ) d t over any range. The estimation of the parameter of maximum likelihood is computed by the likelihood function of evaluation, which is given as:
l ( p , q | t ) = n l o g ( q p )

3.2. The Sobol Sequence

Russian mathematician named Sobol carried out the Sobol distribution in 1967 [50] to reconstruct coordinates. For each dimension d z , coordinate contains the relation of linear recurrences, and for the non-negative instance a z , the binary expression for liner recurrence can be defined as:
a = a 1 2 0 + a 2 2 1 + a 3 2 2 + + a z 2 z 1
For dimension d z , the instance i th can be generated using Equation (4):
x i D = i 1 v 1 D + i 2 v 2 D + + i z v z D
where v 1 D denotes the binary function of the k th direction of an instance v i D at the dimension d z , v i D can be calculated using Equation (5):
V k D = c 1 v k 1 D + c 2 v k 2 D + + c z v z 1 D + ( v i z D 2 z )
c z describes the polynomial coefficient where k > z .

3.3. The Halton Sequence

Halton sequence, designed as an improved version of Van der Corput sequence, was proposed by the authors in [51]. Halton sequences use a base of c o p r i m e to generate random points. The Algorithm 1 pseudocode to generate Halton sequences is as follows:
Algorithm 1 Halton Sequences
Halton ():
// input: Size = z and base = b c m with Dimension = d
// output: population instances = p
Fix the interval over
m a x 1
m i n 0
For each iteration ( k 1 , k 2 , k 3 k z ) : do
For each particle { p 1 , p 2 , p 3 p z }
m a x = m a x / b c m
m i n = m i n + m a x z   m o d   b c m
z = z / b c m

3.4. The Well Sequence

Well-equidistributed long-period linear (WELL) sequence was proposed by [52]. Initially, it carried out an updated version of the Mersenne twister algorithm. The Algorithm 2 for generating WELL distribution is given as:
Algorithm 2 WELL Sequences
WELL ():
t 0 = ( m x & v k , r 1 ) + ( m x & v k , r 2 )
t 1 = ( A 0 v k , 0 ) + ( A 1 v k , m 1 )
t 2 = ( A 2 v k , m 2 ) + ( A 3 v k , m 3 )
t 3 = t 2 + t 1
t 4 = t 0 A 4 + t 1 A 5 + t 2 A 6 + t 3 A 7
v k + 1 , r 1 = v k , r 2   &   m x
for i r 2 . . 2   d o      v k + 1 , i = v k , i 1
v k + 1 , 1 = t 3
v k + 1 , 0 = t 4
Return y k = v k , 0
The algorithm stated above describes the general recurrence for the WELL distribution. The description for the algorithm is: x and r two integers with the interval of r > 0   a n d   0 < x < k and k = r w x , w is the weight factor of distribution. A 0 A 7 represents the binary matrix of size r w having r bit block m x , describing the bitmask that holds the first w x bits. t 0 t 7 are the temporary vector variables.

3.5. The Knuth Sequence

As discussed above, inbuilt library function is used, K n u t h ( x m i n , x m a x ) to generate Knuth sequences of random points. Knuth sequence is designed and was proposed by the authors in [53]. The pseudocode to generate Knuth sequences is as follows:
-To shuffle an array an of n elements (indices 0 … n − 1):
for i from 0 to n − 2 do
j ← random integer such that i ≤ j < n
exchange a[i] and a[j]

3.6. The Torus Sequence

Torus is a geometric term that was firstly used by the authors in [54] to generate a Torus mesh for a geometric coordinate system. In-game development Torus mesh is commonly used and can be created using a left-hand coordinate system or right-hand coordinate system. The shape for the Torus at 1D, 2D, and 3D is a circle and 2D rectangle, respectively. The Torus in 3D can be represented by the following Equations (6)–(8).
a ( θ , δ ) = ( D + r cos θ ) cos δ
b ( θ , δ ) = ( D + r cos θ ) sin δ
c ( θ , δ ) = r sin δ
where their angles of circles are θ, δ and D is the distance from tube center to Torus center r denotes to the radius of circle. Inspired by this mesh with Torus, a low-discrepancy sequences have been generated that being initialize with the prime series as Torus effect. Equation (9) shows the mathematical notation for Torus series.
a k = ( f ( k s 1 ) , , f ( k s d ) )
where s1 denotes the series of ith prime number, and f is a fraction which can be calculated by f = a − floor(a). Due to the prime constraints, the dimension for the Torus is limited to the 100,000, only if we use parameter prime in Torus function. For more than 100,000 dimensions the number must be provided manually.
In Figure 1, Figure 2, Figure 3, Figure 4, Figure 5 and Figure 6 random points following the Uniform, Sobol, Halton, WELL, Knuth, and Torus distribution are shown by bubble plot in which the y-axis represents the random values and the x-axis shows the relevant index of the concerned point in the table.
We have made our first significant contribution to this study by introducing three novel methods of initialization of population WE-PSO, KN-PSO, and TO-PSO.
Algorithm 3 shows the proposed distribution-based PSO initialization:
Algorithm 3 Proposed Pseudocode of PSO Using Novel Method of Initialization
  • Initialize the swarm
  • Set epoch count I = 0 , population size N z , Dimension of the problem D z , w m a x and w m i n
  • For each particle P z
  • Initialize x z , as x z =   WELL ,   Knuth ,   Torus   ( x m i n , x m a x )
  • Initialize the Particle velocity as, v z = R a n d ( x m i n , x m a x )
  • Compute the fitness score f z
  • Set global best position g z b e s t as m a x ( f 1 , f 2 , f 3 . . f z ) where f z g l o b a l l y   o p t i m a l   f i t n e s s
  • Set local best position p z b e s t as m a x ( f 1 , f 2 , f 3 . . f z ) where f z l o c a l l y   o p t i m a l   f i t n e s s
  • Compare the current particle’s fitness score x z in the swarm and its old local best location p z b e s t  If the current fitness score x z is greater than p z b e s t , then substitute p z b e s t , with   x z ; else retain the x z unchanged
  • Compare the current particle’s fitness score x z in the swarm and its old global best location g z b e s t  If the current fitness score x z is greater than g z b e s t , then substitute g z b e s t , with   x z ; else retain the x z unchanged
  • Compute   v z + 1 → updated velocity vector
  • Compute   x z + 1 → updated position vector
  • Go to step 2; If the stopping criteria does not met; else terminat.
In our other contribution in the paper, we proposed by introducing three new methods of initialization of population DE-KE, DE-WE, and DE-TO.
Algorithm 4 shows the proposed distribution-based DE initialization.
Algorithm 4 Proposed Pseudo Code of DE Using Novel Method of Initialization
Input: 𝑥𝑖 = (𝑥𝑖,1, 𝑥𝑖,2, 𝑥𝑖,3, …, 𝑥𝑖,𝐷), Population size ‘N-P’, Problem Size ‘D’, Mutation Rate ‘F’, Crossover Rate ‘C-R’; Stopping Criteria {Number of Generation, Target}, Upper Bound ‘U’, Lower Bound ‘L’
Output: 𝑥𝑖, = Global fitness vector with minimal fitness value
Pop = Initialize of Paraments (N-P, D, U, L);
Generate initial population Using WELL,Knuth,Torus
While (Stopping Criteria ≠ True) do
Best Vector = Evaluate Pop (Pop);
vx = Select Rand Vector (Pop);
I = Find Index Vector (vx);
Select Rand Vector (Pop,v1,v2,v3) where v1 ≠ v2 ≠ v3 ≠ vx
vy = v1, + F(v2−v3)
For (i = 0; i++; i < D−1)
If (randj [0, 1) < C-R) Then
U[i] = vx [i].
else
U[i] = vy [i]
End For loop
If (Cost Fun Vector(U) ≤ Cost Fun Vector (vx)) Then
Update Pop (U, I, Pop);
End IF
End While
Retune Best Vector

4. Experimental Setup

The parameter for the fair performance comparison is presented in Table 1. The algorithm settings are shown in Table 2. For the impartial progress, the correlation condition of the performance evaluation alongside the proposed strategy, all of the parameter was set to the uniform.
The simulation has been simulated in C++ and applied on a computer using the C++ language on the computer with the Windows 10 operating system with the specifications of 8 gigabytes of ram and 2.3 GHz Core (M) 2 Duo CPU processor. Hence in our study, we used it to examine the optimization outcomes of the quasirandom-based approaches. A list of those functions is available in Table 3. In Table 3, D shows the dimensionality of the problem, S represents the interval of the variables, and fmin denotes the standard global optimum minimum value. In the parameters for the simulation used as c1 = c2 = 1.45, inertia weight w is used in the interval [0.9, 0.4], and swarm size is 50. For simulation, the function dimensions are D = 10, 20, and 30, and the maximum number of epochs is 3000. All techniques have been applied to similar parameters for comparatively effective results. In order to check the performance of each quasirandom sequence bases approach, all of them were tested for 10 runs.

5. Simulation Results and Discussion

5.1. Results and Graphs on PSO Approaches

The comparative analysis can be seen from Table 4 that with a smaller dimension size, standard PSO performs well (D = 10), but as the size of dimension increases, KN-PSO outperforms in convergence significantly.

5.2. Friedman and Kruskal–Wallis Test on PSO Approaches

Mean ranks obtained by Kruskal–Wallis and Friedman test are given in the Table 5.

5.3. Discussion on PSO Results

To measure the execution of proposed approaches WELL-based PSO (WE-PSO), Torus-based PSO (TO-PSO), and Knuth-based PSO (KN-PSO), on a group of 16 nonlinear benchmark test functions, have been utilized to make the comparison of WE-PSO, TO-PSO, and KN-PSO with standard PSO, SO-PSO, and H-PSO. These functions are generally applied to investigate the performance of any technique. Hence in our study, we used it to examine the optimization outcomes of the quasirandom-based approach of WE-PSO, TO-PSO, KN-PSO, SO-PSO, and H-PSO.

A. Discussion

The purpose of this study continues to observe whereby the unique characteristics of experimental results rely on dimensions of these standard benchmark functions. In the experiments, three simulation experiments were performed, where the following features of the quasirandom sequence were observed:
  • Effect of using different initializing PSO approaches
  • Effect of using different dimensions for problems
  • A comparative analysis
The objective of this study to find the most suitable initializing method for the PSO, and during the first experiment, the proposed WE-PSO, TO-PSO, and KN-PSO with other approached SO-PSO, H-PSO, and standard PSO were investigated. The objective of second simulation is to find the nature of dimension regarding standard function optimization. Lastly, the simulation results of WE-PSO, TO-PSO, and KN-PSO were compared with standard PSO, SO-PSO, and H-PSO. In the rest of the paper, simulation results were discussed in detail.
Figure 6, Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13, Figure 14, Figure 15, Figure 16, Figure 17, Figure 18, Figure 19, Figure 20, Figure 21 and Figure 22 contain the graphical representation of the comparisons of proposed WE-PSO, TO-PSO, and KN-PSO with standard PSO, H-PSO, and SO-PSO. In the x-axis dimensions of the problem, 10, 20, and 30 are presented, while the y-axis represens the mean best against each dimension of the problem.
We can see that majority of the figures contains better convergence curve for KN-PSO on functions F1, F2, F3, F4, F5, F6, F17, F8, F9, F10, F11, F12, F13, F14, and F15 over WE-PSO, TO-PSO, H-PSO, SO-PSO, and standard PSO on all dimensions comprehensively. The other proposed approach TO-PSO provides better results over WE-PSO on functions F1, F2, F4, F6, F17, F8, F9, F10, F14, F15, and F16 and beats on all functions for PSO, SO-PSO, and TO-PSO.
i.
Effect of Using Different Initializing PSO Approaches
In this simulation, PSO is initialized with the WELL sequence (WE-PSO), Torus Sequence (TO-PSO), and Knuth Sequence (KN-PSO) instead of uniform distribution. The variant proposed WE-PSO, TO-PSO, and KN-PSO is compared with other initialized approaches Sobol sequence (SO-PSO), Halton Sequence (H-PSO), and standard PSO. The experimental results give superior results in higher dimensions for KN-PSO on other SO-PSO, H-PSO, PSO, and proposed approach TO-PSO and WE-PSO.
ii.
Effect of Using Different Dimensions for Problems
The core objective of this simulation setup is to find the superiority of results depending upon the dimension of the functions that are to be optimized. In experiments, three dimensions for benchmark functions D = 10, D = 20, and D = 30 were used. Simulation results were presented in Table 2. From these simulation results, it was found that functions having larger dimensions found tougher to optimize and it can be seen from the Table 2 when dimension size is D = 20 and D = 30, and our proposed approach KN-PSO shows belter result on higher dimensions than other approaches such as WE-PSO, TO-PSO, standard PSO, H-PSO, and SO-PSO.
iii.
Comparative Analysis
KN-PSO is compared with the other approaches such as WE-PSO, TO-PSO, SO-PSO, H-PSO and standard PSO, where each technique true value is presented for comparison with other techniques for the same nature of problems. Standard benchmark functions are presented in the Table 3, and their parameter settings are also shown in the Table 2. The Table 4 shows that with dimension D-30, KN-PSO is more superior and outperforms in convergence than the WE-PSO, TO-PSO, standard PSO, SO-PSO, and H-PSO. The comparative analysis can be seen from Table 4 that with a smaller dimension size, standard PSO performs well (D = 10), but as the size of dimension increases, KN-PSO outperforms in convergence significantly. Hence, KN-PSO is best for higher dimensions. The experimental results from Table 4 show that KN-PSO outclasses the results of WE-PSO, TO-PSO, SO-PSO, H-PSO, and traditional PSO for all functions. It can be seen that the TO-PSO outperformed the results of the other techniques in all functions on SO-PSO, H-PSO, and standard PSO, while the other approach H-PSO performed better on functions f4, f1, f2 for 20D. However, H-PSO gives an overall poor result on higher dimensions, and SO-PSO gives a slightly better result on functions f8, f9, f15 on 10-D but the worst result on larger dimensions). Standard PSO did not provide a better result. Figures from Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13, Figure 14 and Figure 15 depict that WE-PSO outperforms in simulation results compared to other approaches for solving the standard benchmark tests functions for dim size D = 10, D = 20, and D = 30.
For the validation of the numerical results, mean ranks obtained by Kruskal–Wallis and Friedman test for KN-PSO, WE-PSO, TO-PSO, SO-PSO, HA-PSO, and standard PSO are given in the Table 5.

5.4. Discussion on DE Results

Population initialization is a vital factor in evolutionary computing-based algorithm, which considerably influences the diversity and convergence. Rather than applying the random distribution for initialization, quasirandom sequences are used to initialize the population. In this paper, the capability of DE has been extended to make it suitable for the optimization problem by introducing, new initialization techniques Knuth sequence-based DE (DE-KN), the Torus-based sequence-based (DE-TO) and the WELL sequence-based DE (DE-WE) to solve the optimization problems in large dimension search spaces. For global optimization, the most considerable variety of benchmark problems can be used. All benchmark problems have their own individual abilities, and the variety of detailed characteristics of such functions explains the level of complexity for benchmark problems. For the efficiency analysis of the abovementioned optimization algorithms, Table 3 displays the benchmark problems that are utilized. The Table 2 explains the following contents of benchmark problems: name, range, domain, and formulas. In this study, those benchmark problems are incorporated and have been extensively utilized in the literature for conveying a deep knowledge of the performance related to the abovementioned optimization algorithms. To measure the effectiveness and robustness of optimization algorithms, benchmark functions are applied. In this study, 16 computationally expensive black-box functions are applied to their various abilities and traits. The purpose of utilizing these benchmark functions is to examine the effectiveness of the abovementioned proposed approaches.
In this section, for a comparison among low-discrepancies sequence, methods are performed with each other with reference to capabilities and efficiency with the help of high-dimensional 15 benchmark functions. Nevertheless, the whole performance of optimization algorithms varies on the basis of setting parameters and also with other testing criteria. Benchmark problems may be embedded to demonstrate the performance of the low-discrepancies sequence approaches, at various complex levels. Table 6 contains the experimental simulation results on benchmark functions. The exhaustive statistical results are explained in Table 7. From Figure 23, Figure 24, Figure 25, Figure 26, Figure 27, Figure 28, Figure 29, Figure 30, Figure 31, Figure 32, Figure 33, Figure 34, Figure 35, Figure 36, Figure 37 and Figure 38, the experimental results of constrained benchmark test functions are only exhibited by having a surface with D = 10, 20, and 30. The experimental results of this work may not contemplate the entire competency of new proposed low-discrepancies sequence in accordance with the all possible conditions The core objective of this section is to review the consequences of tested optimization approaches in high dimensionality, regarding to the accuracy and reliability of achieved solutions at the time of solving complex and computationally expensive optimization problems. From Figure 23, Figure 24, Figure 25, Figure 26, Figure 27, Figure 28, Figure 29, Figure 30, Figure 31, Figure 32, Figure 33, Figure 34, Figure 35, Figure 36, Figure 37 and Figure 38, the performances of following methods: DE-KN, DE-WE, DE-TO, DE-S, DE-H, and traditional DE, are compared for 16 benchmark functions. In the graphs, the horizontal axis displays the total number of iterations, while on the other hand, the vertical axis displays the mean value of objective functions at the fixed number of function computations. Correspondingly, the value achieved in each iteration is operated as a performance measure. As a result, the exploitation ability of traditional DE is moderately low, particularly for high-dimensional problems. The results are also disclosed that traditional DE, DE-SO, and DE-H are only effective in performance while they are tackling with expensive design problems having low dimensionality.
Beside this, H-DE has excellent control on the high-dimensionality problems compared to other methods in spite of complexity and the superficial topology of the examined problems. Figure 23, Figure 24, Figure 25, Figure 26, Figure 27, Figure 28, Figure 29, Figure 30, Figure 31, Figure 32, Figure 33, Figure 34, Figure 35, Figure 36, Figure 37 and Figure 38 shows the achievements of traditional DE, DE-S, DE-WE, DE-KN, and DE-TO algorithms with regard to their efficiency and capability. The results are demonstrated that DE-H outperforms in higher dimensionality problems. By summarizing it, the dimensionality strongly influences the working of most algorithms. However, it is observed that DE-H is more consistent during the increment of dimensions of the problem. Due to this consistency of DE-H, it is proven that the DE-H algorithm has the greater capability of exploration. For statistical comparison, widely known mean ranks obtained by Kruskal–Wallis and Friedman test, which are implemented to compare the implications between the DE-H algorithm and other algorithms in DE-KN, DE-WE, DE-S, DE-TO, and standard DE, are given in the Table 5.

5.5. Results and Graphs on DE Approaches

The experimental simulation results on benchmark functions are shown in Table 6.
Table 6. Comparative results for all DE-based approaches on 16 standard benchmark functions.
Table 6. Comparative results for all DE-based approaches on 16 standard benchmark functions.
FunctionsDIM × IterDEDE-HDE-SDE-TODE-WEDE-KN
F110 × 10001.1464 × 10−442.1338 × 10−445.8561 × 10−447.4117 × 10−457.4827 × 10−395.7658 × 10−39
20 × 20003.3550 × 10−467.2338 × 10−461.3545 × 10−451.2426 × 10−459.6318 × 10−457.1501 × 10−45
30 × 30008.8946 × 10−471.2273 × 10−459.4228 × 10−461.6213 × 10−466.2007 × 10−465.7425 × 10−46
F210 × 10000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+00
20 × 20000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+00
30 × 30001.8392 × 10+011.1846 × 10+011.8871 × 10+013.7132 × 10−015.0821 × 10+006.6313 × 10+00
F310 × 10005.00325 × 10−441.5019 × 10−389.3956 × 10−444.7807 × 10−441.6251 × 10−381.3411 × 10−38
20 × 20002.56987 × 10−454.1485 × 10−441.5339 × 10−443.0262 × 10−459.5984 × 10−441.3606 × 10−43
30 × 30001.01692 × 10−452.7349 × 10−454.0581 × 10−454.5726 × 10−454.5686 × 10−455.4659 × 10−45
F410 × 10005.81825 × 10−423.0950 × 10−362.2300 × 10−411.6903 × 10−411.1331 × 10−363.8869 × 10−36
20 × 20002.70747 × 10−431.0658 × 10−411.6730 × 10−421.3490 × 10−421.3094 × 10−416.0053 × 10−42
30 × 30002.99887 × 10−431.4032 × 10−424.4442 × 10−425.9186 × 10−434.6922 × 10−431.4829 × 10−42
F510 × 10001.65318 × 10−434.7939 × 10−387.0329 × 10−434.8106 × 10−434.3219 × 10−383.5770 × 10−38
20 × 20001.39082 × 10−443.6325 × 10−434.2191 × 10−442.7448 × 10−445.8557 × 10−431.4008 × 10−43
30 × 30006.07162 × 10−451.7557 × 10−441.6295 × 10−442.0582 × 10−448.6773 × 10−454.2285 × 10−44
F610 × 10007.8201 × 10−963.8819 × 10−969.7956 × 10−962.3292 × 10−958.4774 × 10−942.8037 × 10−95
20 × 20001.6847 × 10−1258.6880 × 10−1245.9005 × 10−1228.7800 × 10−1233.7438 × 10−1241.3947 × 10−124
30 × 30002.4533 × 10−1401.5487 × 10−1395.7211 × 10−1384.4492 × 10−1376.5749 × 10−1403.4442 × 10−137
F710 × 10008.0217 × 10−757.3243 × 10−675.7807 × 10−661.0243 × 10−731.9035 × 10−671.4359 × 10−65
20 × 20004.0682 × 10−711.5037 × 10−701.5747 × 10−691.0623 × 10−705.5546 × 10−702.3507 × 10−70
30 × 30008.5895 × 10−686.6009 × 10−683.3919 × 10−672.6036 × 10−671.1587 × 10−672.1901 × 10−67
F810 × 10007.0221 × 10−1203.4271 × 10−1082.7718 × 10−1086.3092 × 10−1183.9423 × 10−1069.9394 × 10−108
20 × 20005.2096 × 10−1087.7158 × 10−891.4732 × 10−1068.8720 × 10−1073.4490 × 10−1072.2539 × 10−106
30 × 30001.2538 × 10−981.8071 × 10−981.1085 × 10−957.2462 × 10−982.5375 × 10−995.8040 × 10−98
F910 × 10000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+00
20 × 20000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+00
30 × 30000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+00
F1010 × 10001.3459 × 10−752.6493 × 10−662.6884 × 10−663.6168 × 10−673.8397 × 10−671.8408 × 10−66
20 × 20003.0478 × 10−711.6106 × 10−695.5253 × 10−692.7746 × 10−705.3662 × 10−705.0931 × 10−70
30 × 30008.2514 × 10−681.0937 × 10−664.1120 × 10−671.3055 × 10−673.5397 × 10−686.0736 × 10−67
F1110 × 10002.3417 × 10−421.2483 × 10−411.3726 × 10−416.3337 × 10−424.8161 × 10−427.3464 × 10−42
20 × 20008.4769 × 10−443.5140 × 10−433.3777 × 10−432.4721 × 10−431.9553 × 10−433.6961 × 10−43
30 × 30003.6888 × 10−446.9938 × 10−442.5123 × 10−431.4710 × 10−434.0019 × 10−443.9503 × 10−43
F1210 × 10002.3304 × 10+004.4354 × 10+003.4520 × 10+005.1229 × 10+003.8782 × 10+002.7840 × 10+00
20 × 20003.1768 × 10+043.9596 × 10+043.8814 × 10+042.9488 × 10+044.1181 × 10+044.0914 × 10+04
30 × 30001.1760 × 10+061.0300 × 10+061.3402 × 10+061.2008 × 10+061.0916 × 10+061.0160 × 10+06
F1310 × 10001.3940 × 10−651.3756 × 10−643.1956 × 10−669.3609 × 10−645.4864 × 10−639.2695 × 10−63
20 × 20002.0163 × 10−1118.5333 × 10−1108.5260 × 10−1113.9836 × 10−1095.0102 × 10−1154.4624 × 10−110
30 × 30001.4146 × 10−1564.3434 × 10−1564.4702 × 10−1544.3862 × 10−1511.0781 × 10−1531.0142 × 10−149
F1410 × 10009.1259 × 10−242.1900 × 10−232.5559 × 10−232.9039 × 10−231.9174 × 10−233.3427 × 10−23
20 × 20002.6867 × 10−253.8631 × 10−251.5177 × 10−245.5714 × 10−254.5049 × 10−255.6503 × 10−25
30 × 30005.9241 × 10−268.6401 × 10−268.4348 × 10−261.4630 × 10−259.7932 × 10−261.4921 × 10−25
F1510 × 10001.0493 × 10−1854.0276 × 10−1815.0331 × 10−1823.1770 × 10−1831.1698 × 10−1802.6563 × 10−182
20 × 20002.9407 × 10−1599.9152 × 10−1592.1401 × 10−1589.0345 × 10−1563.8871 × 10−1588.0144 × 10−160
30 × 30004.6769 × 10−1381.0737 × 10−1377.0544 × 10−1388.0376 × 10−1384.9091 × 10−1391.1054 × 10−137
F1610 × 10001.8635 × 10−041.8109 × 10−024.9798 × 10−025.8605 × 10−041.4858 × 10−023.7220 × 10−02
20 × 20001.1032 × 10+001.6605 × 10+001.7157 × 10+001.4875 × 10+001.5697 × 10+001.2008 × 10+00
30 × 30002.8283 × 10+012.2049 × 10+012.9388 × 10+012.8205 × 10+012.5794 × 10+012.9526 × 10+01

5.6. Friedman and Kruskal–Wallis Test on DE Approaches

In Table 7, the exhaustive statistical results are explained.
Table 7. Mean ranks obtained by Kruskal–Wallis and Friedman test for all DE-based approaches.
Table 7. Mean ranks obtained by Kruskal–Wallis and Friedman test for all DE-based approaches.
Friedman ValueKruskal-Wallis
DE63.7465.11
DE-H59.3160.41
DE-S64.0165.05
DE-TO63.7665.35
DE-WE63.3563.93
DE-KN63.3364.06

6. Comparison of PSO and DE Regarding Data Classification

6.1. NN Classifications with PSO-Based Initialization Approaches

For further verification of performance of proposed algorithms TO-PSO, WE-PSO and KN-PSO, a comparative study for real-world benchmark data sets problem is tested for training of neural network. We have performed experiments using seven benchmark data sets (diabetes, heart, wine, seed, vertebral, blood tissue, and mammography) exerted from the world-famous machine-learning repository of University of California, Irvine (UCI). Training weights are initialized within interval [−50, 50]. Accuracy of feed-forward neural network is tested in the form of root mean squared error (RMSE). Table 8 shows the characteristics of the data sets used.

Discussion

Multilayer feed-forward neural network is trained with a backpropagation algorithm, standard PSO, SO-PSO, and H-PSO, and proposed TO-PSO, KN-PSO, and WE-PSO. Comparison of these training approaches was tested on real classification problem data sets taken from UCI repository. The crossvalidation method was used to compare the performances of different classification techniques. In the paper, k-fold crossvalidation method used for comparison of classification performances for the training of neural network with back propagation, standard PSO, SO-PSO, and H-PSO and proposed TO-PSO, KN-PSO, and WE-PSO is used. The k-fold crossvalidation method was proposed and used in the experimental with value k = 10. Data set was divided into 10 chunks, such that each chunk of data contains the same proportion of each class of data set. One chunk is used for testing, while nine chunks are used for the training phase. The experimental results of algorithms such as with backpropagation, standard PSO, SOPSO, and H-PSO and proposed TO-PSO, KN-PSO, and WE-PSO are compared with each other on seven well-known real data sets taken from UCI, and their performances are evaluated. In Table 9, the simulation results show that the training of neural network with KN-PSO algorithm out performs in accuracy and is capable of providing a better classification accuracy than the other traditional approaches. KN-PSO algorithm may be used effectively for data classification and statistical problems in the future as well. Figure 39 represents the accuracy graph for seven data sets.
The classification testing accuracy were imported from Microsoft Excel Spreadsheet to the software RStudio version 1.2.5001 (Boston, MA, USA) to get assurance of the winner approach statistically among all the other approaches. The testing accuracy of all seven variants of PSONN was analyzed by one-way ANOVA test and post hoc Tukey’s multicomparison test with a 0.05 significance level [55]. Table 10 depicts the results of one-way ANOVA of the testing accuracy of classification data. The significance value in Table 10 is 0.04639, which is less than 0.05, giving evidence that there is a significant difference among all variants of PSONN with a 95% confidence level. According to this, the variants of PSONN are significantly distinct from each other. Figure 40 represents the graph of one-way ANOVA results, which conclude that KN-PSONN significantly outperforms all other variants of PSONN. Figure 41 represents the results of multicomparisons of PSONN variants through post hoc Tukey’s test. The resultant graph depicts that KN-PSONN variant is significantly different from all other variants. According to the results of Figure 39, KN-PSONN is proved statistically different from all other approaches of PSONN with 95% confidence level.

6.2. NN Classifications with DE-Based Initialization Approaches

The proposed approaches DE-KN, DE-TO, and DE-WE and family of low-discrepancy sequences are extremely suitable for tackling the global optimization problems. A comparative study for real-world benchmark data sets problem is tested for the training of neural network. We have performed experiments using seven benchmark data sets (diabetes, heart, wine, seed, vertebral, blood tissue, and mammography) exerted from the world-famous machine-learning repository of UCI. Training weights are initialized within the interval [−50, 50]. Accuracy of the feed-forward neural network is tested in the form of root mean squared error (RMSE)

Discussion

Multilayer feed-forward neural network is trained with a backpropagation algorithm, standard DE, DE-S, and DE-H and proposed DE-TO, DE-KN, and DE-WE. For this goal, we have prepared the multilayer feed-forward neural network utilizing the process of weight optimization. The performance of the DE, DE-S, DE-H, DE-TO, DE-KN, and DE-WE and state-of-the-art NN algorithms are examined on 10 well-known data sets which have been taken directly from the worldwide UCI repository of machine learning. The features of those informational indexes are given in Table 8. These features include the total units participated against each data set, the number of total input instances, the data set nature, and the number of classes against each data set, i.e., binary class problem or multiclass problem. The impact of increasing the number of target classes is independent, as the proposed strategy is purely concerned with weight optimization rather thanfeature selection or reducing high dimensionality. In addition, 10-fold cross-validation method has been carried out for the training and testing process. The experimental results of algorithms such as with backpropagation, standard DE, DE-S, DE-H, DE-WE, DE-TO, and DE-KN are compared with each other on seven well-known real data sets taken from UCI, and their performances are evaluated. In Table 11, the simulation results show that training of neural networks with DE-H algorithm outperforms in accuracy and is capable of providing better classification accuracy than the other traditional approaches. DE-H algorithm may be used effectively for data classification and statistical problems in the future as well. Figure 42 represents the accuracy graph for seven data sets
To prove the experimental results statistically, the testing accuracy of classification data sets were loaded to the software RStudio (1.2.5001 version). The classification results of seven approaches of DE were tested by the one-way ANOVA statistical test and post hoc Tukey’s pair-wise comparison statistical test using significance level 0.05 [19,55]. The findings of classification data set with one-way ANOVA are illustrated in Table 12, where the significance = 0.02043, which is less than the abovementioned threshold of significance level. The findings in Table 12 proved that there are significant dissimilarities in all variants of DE with 95% confidence level. Figure 43 demonstrates the graph of one-way ANOVA that gives the evidence that H-DE is significantly better than other approaches of DE. Figure 44 models the findings of pair-wise comparisons of DE approaches with post hoc Tukey’s statistical test. The simulated graph describes that H-DE approach is statistically significantly dissimilar as compared to other approaches of DE having 95% confidence level.

7. Conclusions

This paper presents three novel pseudorandom initialization strategies called WELL sequence, Knuth sequence, and Torus sequence, which are used to initialize the population in the search space for PSO and DE algorithm. The experimental validation of proposed approaches by using the family of low-discrepancy sequences is tested on the comprehensive set of benchmark test functions and training of the artificial neural network. The simulation results reveal that the use of the family of low-discrepancy sequences maintains the diversity of the swarm, improves the convergence speed, and finds the better region of the swarm. The proposed families of low-discrepancy sequences contain higher diversity and enhance the local searching ability. The experimental results depict that the KN-PSO and H-DE have superior accuracy of convergence and avoid local optima in a better way. The proposed techniques are compared with a family of low-discrepancy sequences approaches for PSO and DE and also compared with traditional PSO and DE using random distribution and provide better results. The core objective of this research is applicable to other stochastic-based metaheuristic algorithms which develop the future direction of our work.

Author Contributions

Data curation, M.R.H.; Formal analysis, A.A.B.A.I.; Methodology, K.N.; Project administration, J.J.P.C.R.; Resources, D.B.R.; Writing—review & editing, W.H.B. All authors have read and agreed to the published version of the manuscript.

Funding

The manuscript APC is supported by Universiti Malaysia Sabah, Jalan UMS, 88400, KK, Malaysia. Furthermore, this work is partially funded by FCT/MCTES through national funds and when applicable co-funded EU funds under the Project UIDB/50008/2020; and by Brazilian National Council for Scientific and Technological Development - CNPq, via Grant No. 313036/2020-9.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yang, C.H.; Chang, H.W.; Ho, C.H.; Chou, Y.C.; Chuang, L.Y. Conserved PCR primer set designing for closely-related species to complete mitochondrial genome sequencing using a sliding window-based PSO algorithm. PLoS ONE 2011, 6, e17729. [Google Scholar] [CrossRef] [Green Version]
  2. Mahi, M.; Baykan, Ö.K.; Kodaz, H. A new hybrid method based on particle swarm optimization, ant colony optimization and 3-opt algorithms for traveling salesman problem. Appl. Soft Comput. 2015, 30, 484–490. [Google Scholar] [CrossRef]
  3. Rao, S.S. Engineering Optimization: Theory and Practice; John Wiley & Sons: Hoboken, NJ, USA, 2019. [Google Scholar]
  4. Zhang, G.; Lu, J.; Gao, Y. Multi-Level Decision Making: Models, Methods and Applications. 2015. Available online: https://0-www-springer-com.brum.beds.ac.uk/gp/book/9783662460580 (accessed on 15 April 2021).
  5. Beni, G.; Wang, J. Swarm Intelligence in Cellular Robotic Systems, in Robots and Biological Systems: Towards a New Bionics? Springer: New York, NY, USA, 1993; pp. 703–712. [Google Scholar]
  6. Acharya, J.; Mehta, M.; Saini, B. Particle swarm optimization based load balancing in cloud computing. In Proceedings of the 2016 International Conference on Communication and Electronics Systems (ICCES), Coimbatore, India, 21–22 October 2016. [Google Scholar]
  7. Zhang, Y.; Xiong, X.; Zhang, Q.D. An improved self-adaptive PSO algorithm with detection function for multimodal function optimization problems. Math. Probl. Eng. 2013, 2013, 1–8. Available online: https://econpapers.repec.org/article/hinjnlmpe/716952.htm (accessed on 15 April 2021). [CrossRef]
  8. Eberhart, R.; Kennedy, J. Particle swarm optimization. In Proceedings of the IEEE International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; pp. 1942–1948. [Google Scholar]
  9. Yang, X.-S. A new metaheuristic bat-inspired algorithm. In Nature Inspired Cooperative Strategies for Optimization (NICSO 2010); Springer: New York, NY, USA, 2010; pp. 65–74. [Google Scholar]
  10. Dorigo, M.; Di Caro, G. Ant colony optimization: A new meta-heuristic. In Proceedings of the 1999 Congress on Evolutionary Computation—CEC99, Washington, DC, USA, 6–9 July 1999; pp. 1470–1477, Cat. No. 99TH8406. [Google Scholar]
  11. Pham, D.T.; Ghanbarzadeh, A.; Koç, E.; Otri, S.; Rahim, S.; Zaidi, M. The bees algorithm—A novel tool for complex optimisation problems. In Intelligent Production Machines and Systems; Elsevier: Amsterdam, The Netherlands, 2006; pp. 454–459. [Google Scholar]
  12. Poli, R.; Kennedy, J.; Blackwell, T. Particle Swarm Optimization; Springer: New York, NY, USA, 2007; Volume 1, pp. 33–57. [Google Scholar]
  13. Bai, Q. Analysis of particle swarm optimization algorithm. Comput. Inf. Sci. 2010, 3, 180. [Google Scholar] [CrossRef] [Green Version]
  14. AlRashidi, M.R.; El-Hawary, M.E. A survey of particle swarm optimization applications in electric power systems. IEEE Trans. Evol. Comput. 2008, 13, 913–918. [Google Scholar] [CrossRef]
  15. Zhu, H.M.; Wu, Y.P. A PSO algorithm with high speed convergence. Control Decis. 2010, 25, 20–24. [Google Scholar]
  16. Chen, B.; Lei, H.; Shen, H.; Liu, Y.; Lu, Y. A hybrid quantum-based PIO algorithm for global numerical optimization. Sci. China Inf. Sci. 2019, 62, 1–12. [Google Scholar] [CrossRef] [Green Version]
  17. Shi, Y. Particle swarm optimization: Developments, applications and resources. In Proceedings of the 2001 Congress on Evolutionary Computation, Washington, DC, USA, 6–9 July 1999; pp. 81–86, Cat. No. 01TH8546. [Google Scholar]
  18. Chen, S.; Montgomery, J. Particle swarm optimization with thresheld convergence. In Proceedings of the 2013 IEEE Congress on Evolutionary Computation, Cancun, Mexico, 20–23 June 2013; pp. 510–516. [Google Scholar]
  19. Alam, M.N.; Das, B.; Pant, V. A comparative study of metaheuristic optimization approaches for directional overcurrent relays coordination. Electr. Power Syst. Res. 2015, 128, 39–52. [Google Scholar] [CrossRef]
  20. Lu, Z.; Hou, Z. Adaptive Mutation PSO Algorithm. Acta Electronca Sinica 3 2004, 32, 417–420. [Google Scholar]
  21. Rashedi, E.; Nezamabadi-Pour, H.; Saryazdi, S. GSA: A Gravitational Search Algorithm; Elsevier: Amsterdam, The Netherlands, 2009; Volume 179, pp. 2232–2248. [Google Scholar]
  22. Li, X.; Zhuang, J.; Wang, S.; Zhang, Y. A particle swarm optimization algorithm based on adaptive periodic mutation. In Proceedings of the 2008 Fourth International Conference on Natural Computation, Jinan, China, 18–20 October 2008; pp. 150–155. [Google Scholar]
  23. Song, M.-P.; Gu, G.-C. Research on particle swarm optimization: A review. In Proceedings of the 2004 International Conference on Machine Learning and Cybernetics, Shanghai, China, 26–29 August 2004; pp. 2236–2241, Cat. No. 04EX826. [Google Scholar]
  24. Maaranen, H.; Miettinen, K.; Penttinen, A. On Initial Populations of a Genetic Algorithm for Continuous Optimization Problems; Springer: New York, NY, USA, 2007; Volume 37, pp. 405–436. [Google Scholar]
  25. Pant, M.; Thangaraj, R.; Grosan, C.; Abraham, A. Improved particle swarm optimization with low-discrepancy sequences. In Proceedings of the 2008 IEEE Congress on Evolutionary Computation (IEEE World Congress on Computational Intelligence), Hong Kong, China, 1–6 June 2008; pp. 3011–3018. [Google Scholar]
  26. Parsopoulos, K.E.; Vrahatis, M.N. Initializing the particle swarm optimizer using the nonlinear simplex method. Adv. Intell. Syst. Fuzzy Syst. Evol. Comput. 2002, 216, 1–6. [Google Scholar]
  27. Richards, M.; Ventura, D. Choosing a starting configuration for particle swarm optimization. Neural Netw. 2004, 25, 2309–2312. [Google Scholar]
  28. Nguyen, X.H.; Nguyen, Q.U.; McKay, R.I. PSO with randomized low-discrepancy sequences. In Proceedings of the 9th Annual Conference on Genetic and Evolutionary Computation—GECCO ’07, New York, NY, USA, 7–11 July 2007; p. 173. [Google Scholar]
  29. Uy, N.Q.; Hoai, N.X.; McKay, R.I.; Tuan, P.M. Initialising PSO with randomised low-discrepancy sequences: The comparative results. In Proceedings of the 2007 IEEE Congress on Evolutionary Computation, Singapore, 25–28 September 2007; pp. 1985–1992. [Google Scholar]
  30. Thangaraj, R.; Pant, M.; Deep, K. Initializing PSO with probability distributions and low-discrepancy sequences: The comparative results. In Proceedings of the World Congress on Nature & Biologically Inspired Computing (NaBIC), Coimbatore, India, 9–11 December 2009; pp. 1121–1126. [Google Scholar] [CrossRef]
  31. Thangaraj, R.; Pant, M.; Abraham, A.; Badr, Y. Hybrid Evolutionary Algorithm for Solving Global Optimization Problems. In Proceedings of the International Conference on Hybrid Artificial Intelligence Systems, Salamanca, Spain, 10–12 June 2009. [Google Scholar]
  32. Pant, M.; Thangaraj, R.; Singh, V.P.; Abraham, A. Particle Swarm Optimization Using Sobol Mutation. In Proceedings of the 2008 First International Conference on Emerging Trends in Engineering and Technology, Nagpur, India, 16–18 July 2008; pp. 367–372. [Google Scholar]
  33. Du, J.; Zhang, F.; Huang, G.; Yang, J. A new initializing mechanism in Particle Swarm Optimization. In Proceedings of the 2011 IEEE International Conference on Computer Science and Automation Engineering, Shanghai, China, 10–12 June 2011; Volume 4, pp. 325–329. [Google Scholar]
  34. Murugan, P. Modified particle swarm optimisation with a novel initialisation for finding optimal solution to the transmission expansion planning problem. IET Gener. Transm. Distrib. 2012, 6, 1132–1142. [Google Scholar] [CrossRef]
  35. Yin, L.; Hu, X.-M.; Zhang, J. Space-based initialization strategy for particle swarm optimization. In Proceedings of the fifteenth Annual Conference Companion on Genetic and Evolutionary Computation Conference Companion—GECCO ’13 Companion, New York, NY, USA, 6–10 July 2013; pp. 19–20. [Google Scholar]
  36. Jensen, B.; Bouhmala, N.; Nordli, T. A Novel Tangent based Framework for Optimizing Continuous Functions. J. Emerg. Trends Comput. Inf. Sci. 2013, 4. [Google Scholar]
  37. Shatnawi, M.; Nasrudin, M.F.; Sahran, S. A new initialization technique in polar coordinates for Particle Swarm Optimization and Polar PSO. Int. J. Adv. Sci. Eng. Inf. Technol. 2017, 7, 242. [Google Scholar] [CrossRef]
  38. Bewoor, L.; Prakash, V.C.; Sapkal, S.U. Evolutionary Hybrid Particle Swarm Optimization Algorithm for Solving NP-Hard No-Wait Flow Shop Scheduling Problems. Algorithms 2017, 10, 121. [Google Scholar] [CrossRef] [Green Version]
  39. Zhang, J.R.; Zhang, J.; Lok, T.M.; Lyu, M.R. A hybrid particle swarm optimization–back-propagation algorithm for feedforward neural network training. Appl. Math. Comput. 2007, 185, 1026–1037. [Google Scholar] [CrossRef]
  40. Carvalho, M.; Ludermir, T.B. Particle swarm optimization of neural network architectures and weights. In Proceedings of the 7th International Conference on Hybrid Intelligent Systems (HIS 2007), Kaiserslautern, Germany, 17–19 September 2007; pp. 336–339. [Google Scholar]
  41. Mohammadi, N.; Mirabedini, S.J. Comparison of particle swarm optimization and backpropagation algorithms for training feed forward neural network. J. Math. Comput. Sci. 2014, 12, 113–123. [Google Scholar] [CrossRef]
  42. Albeahdili, H.M.; Han, T.; Islam, N.E. Hybrid algorithm for the optimization of training convolutional neural network. Int. J. Adv. Comput. Sci. Appl. 2015, 1, 79–85. [Google Scholar]
  43. Gudise, V.G.; Venayagamoorthy, G.K. Simplex differential evolution. Acta Polytech. Hung. 2009, 6, 95–115. [Google Scholar]
  44. Nakib, A.; Daachi, B.; Siarry, P. Hybrid Differential Evolution Using Low-Discrepancy Sequences for Image Segmentation. In Proceedings of the 2012 IEEE 26th International Parallel and Distributed Processing Symposium Workshops & PhD Forum, Shanghai, China, 21–25 May 2012; pp. 634–640. [Google Scholar]
  45. Tang, L.; Zhao, Y.; Liu, J. An Improved Differential Evolution Algorithm for Practical Dynamic Scheduling in Steelmaking-Continuous Casting Production. IEEE Trans. Evol. Comput. 2013, 18, 209–225. [Google Scholar] [CrossRef]
  46. Wang, L.; Zeng, Y.; Chen, T. Back propagation neural network with adaptive differential evolution algorithm for time series forecasting. Expert Syst. Appl. 2015, 42, 855–863. [Google Scholar] [CrossRef]
  47. Hou, Y.; Zhao, L.; Lu, H. Fuzzy neural network optimization and network traffic forecasting based on improved differential evolution. Future Gener. Comput. Syst. 2018, 81, 425–432. [Google Scholar] [CrossRef]
  48. Panigrahi, S.; Bhoi, A.K.; Karali, Y. A modified differential evolution algorithm trained pi-sigma neural network for pattern classification. Int. J. Soft Comput. Eng. 2013, 3, 133–136. [Google Scholar]
  49. Matsumoto, M.; Nishimura, T. Mersenne twister: A 623-dimensionally equidistributed uniform pseudo-random number generator. ACM Trans. Model. Comput. Simul. TOMACS 1998, 8, 3–30. [Google Scholar] [CrossRef] [Green Version]
  50. Sobol’, I. On the distribution of points in a cube and the approximate evaluation of integrals. USSR Comput. Math. Math. Phys. 1967, 7, 86–112. [Google Scholar] [CrossRef]
  51. Halton, J.H. Algorithm 247: Radical-inverse quasi-random point sequence. Commun. ACM 1964, 7, 701–702. [Google Scholar] [CrossRef]
  52. Panneton, F.; L’Ecuyer, P.; Matsumoto, M. Improved long-period generators based on linear recurrences modulo 2. ACM Trans. Math. Softw. 2006, 32, 1–16. [Google Scholar] [CrossRef]
  53. Knuth, D.E. The Art of Computer Programming; Addison-Wesley: Reading, MA, USA, 1973; Volume 2, p. 51. [Google Scholar]
  54. Williams, H.C.; Nikulin, V.V.; Shafarevich, I.R.; Reid, M. Geometries and Groups. Math. Gaz. 1989, 73, 257. [Google Scholar] [CrossRef]
  55. Ulusoy, U. Application of ANOVA to image analysis results of talc particles produced by different milling. Powder Technol. 2008, 188, 133–138. [Google Scholar] [CrossRef]
Figure 1. Population initialization using the uniform distribution.
Figure 1. Population initialization using the uniform distribution.
Applsci 11 07591 g001
Figure 2. Population initialization using Sobol distribution.
Figure 2. Population initialization using Sobol distribution.
Applsci 11 07591 g002
Figure 3. Population initialization using Halton distribution.
Figure 3. Population initialization using Halton distribution.
Applsci 11 07591 g003
Figure 4. Population initialization using WELL distribution.
Figure 4. Population initialization using WELL distribution.
Applsci 11 07591 g004
Figure 5. Population initialization using Knuth distribution.
Figure 5. Population initialization using Knuth distribution.
Applsci 11 07591 g005
Figure 6. Sample data generated using Torus distribution.
Figure 6. Sample data generated using Torus distribution.
Applsci 11 07591 g006
Figure 7. Convergence curve on fn1.
Figure 7. Convergence curve on fn1.
Applsci 11 07591 g007
Figure 8. Convergence curve on fn2.
Figure 8. Convergence curve on fn2.
Applsci 11 07591 g008
Figure 9. Convergence curve on fn3.
Figure 9. Convergence curve on fn3.
Applsci 11 07591 g009
Figure 10. Convergence curve on fn4.
Figure 10. Convergence curve on fn4.
Applsci 11 07591 g010
Figure 11. Convergence curve on fn5.
Figure 11. Convergence curve on fn5.
Applsci 11 07591 g011
Figure 12. Convergence curve on fn6.
Figure 12. Convergence curve on fn6.
Applsci 11 07591 g012
Figure 13. Convergence curve on fn7.
Figure 13. Convergence curve on fn7.
Applsci 11 07591 g013
Figure 14. Convergence Curve on fn8.
Figure 14. Convergence Curve on fn8.
Applsci 11 07591 g014
Figure 15. Convergence curve on fn9.
Figure 15. Convergence curve on fn9.
Applsci 11 07591 g015
Figure 16. Convergence curve on fn10.
Figure 16. Convergence curve on fn10.
Applsci 11 07591 g016
Figure 17. Convergence curve on fn11.
Figure 17. Convergence curve on fn11.
Applsci 11 07591 g017
Figure 18. Convergence curve on fn12.
Figure 18. Convergence curve on fn12.
Applsci 11 07591 g018
Figure 19. Convergence curve on fn13.
Figure 19. Convergence curve on fn13.
Applsci 11 07591 g019
Figure 20. Convergence curve on fn14.
Figure 20. Convergence curve on fn14.
Applsci 11 07591 g020
Figure 21. Convergence curve on fn15.
Figure 21. Convergence curve on fn15.
Applsci 11 07591 g021
Figure 22. Convergence curve on fn16.
Figure 22. Convergence curve on fn16.
Applsci 11 07591 g022
Figure 23. Convergence curve on fn1.
Figure 23. Convergence curve on fn1.
Applsci 11 07591 g023
Figure 24. Convergence curve on fn2.
Figure 24. Convergence curve on fn2.
Applsci 11 07591 g024
Figure 25. Convergence curve on fn3.
Figure 25. Convergence curve on fn3.
Applsci 11 07591 g025
Figure 26. Convergence curve on fn4.
Figure 26. Convergence curve on fn4.
Applsci 11 07591 g026
Figure 27. Convergence curve on fn5.
Figure 27. Convergence curve on fn5.
Applsci 11 07591 g027
Figure 28. Convergence curve on fn6.
Figure 28. Convergence curve on fn6.
Applsci 11 07591 g028
Figure 29. Convergence curve on fn7.
Figure 29. Convergence curve on fn7.
Applsci 11 07591 g029
Figure 30. Convergence curve on fn8.
Figure 30. Convergence curve on fn8.
Applsci 11 07591 g030
Figure 31. Convergence curve on fn9.
Figure 31. Convergence curve on fn9.
Applsci 11 07591 g031
Figure 32. Convergence curve on fn10.
Figure 32. Convergence curve on fn10.
Applsci 11 07591 g032
Figure 33. Convergence curve on fn11.
Figure 33. Convergence curve on fn11.
Applsci 11 07591 g033
Figure 34. Convergence curve on fn12.
Figure 34. Convergence curve on fn12.
Applsci 11 07591 g034
Figure 35. Convergence curve on fn13.
Figure 35. Convergence curve on fn13.
Applsci 11 07591 g035
Figure 36. Convergence curve on fn14.
Figure 36. Convergence curve on fn14.
Applsci 11 07591 g036
Figure 37. Convergence curve on fn15.
Figure 37. Convergence curve on fn15.
Applsci 11 07591 g037
Figure 38. Convergence curve on fn16.
Figure 38. Convergence curve on fn16.
Applsci 11 07591 g038
Figure 39. Classification Testing Accuracy Results.
Figure 39. Classification Testing Accuracy Results.
Applsci 11 07591 g039
Figure 40. Box plot visualization of the results achieved by the training of FFNN for all PSO-based initialization approaches and BPA for given data set of classification problem.
Figure 40. Box plot visualization of the results achieved by the training of FFNN for all PSO-based initialization approaches and BPA for given data set of classification problem.
Applsci 11 07591 g040
Figure 41. Multicomparison post hoc Tukey test graph of all PSO-based initialization approaches and BPA.
Figure 41. Multicomparison post hoc Tukey test graph of all PSO-based initialization approaches and BPA.
Applsci 11 07591 g041
Figure 42. Classification Testing Accuracy Results.
Figure 42. Classification Testing Accuracy Results.
Applsci 11 07591 g042
Figure 43. Boxplot visualization of the results achieved by the training of FFNN for all DE-based initialization approaches and BPA for given data set of classification problem.
Figure 43. Boxplot visualization of the results achieved by the training of FFNN for all DE-based initialization approaches and BPA for given data set of classification problem.
Applsci 11 07591 g043
Figure 44. Multicomparison post hoc Tukey test graph of DE variants.
Figure 44. Multicomparison post hoc Tukey test graph of DE variants.
Applsci 11 07591 g044
Table 1. Experimental setting of parameters.
Table 1. Experimental setting of parameters.
ParameterValue
Search Space[−100, 100]
Dimensions102030
Iterations100020003000
Population size50
Number of Runs 10
Table 2. Parameters setting of algorithms.
Table 2. Parameters setting of algorithms.
AlgorithmParameters
PSOc1 = c2 = 1.49, w = linearly decreasing
DE𝐹 ∈ [0.4, 1], 𝐶𝑅 ∈ 0.6
Table 3. Function table with characteristics.
Table 3. Function table with characteristics.
Sr.#Function NameObjective FunctionSearch SpaceOptimal Value
01Sphere M i n f ( x ) = i = 1 D x i 2 5.12 x i 5.12 0
02Rastrigin M i n f ( x ) = 10 D + i = 1 D [ x i 2 10 cos ( 2 π x ) ] i 5.12 x i 5.12 0
03Axis parallel hyper-ellipsoid M i n f ( x ) = i = 1 D ( i . x i 2 ) 5.12 x i 5.12 0
04Rotated hyper ellipsoid M i n f ( x ) = i = 1 D j = 1 i ( x j 2 ) 65.536 x i 65.536 0
05Moved Axis M i n f ( x ) = i = 1 D 5 i . x i 2 5.12 x i 5.12 0
06Sum of different power M i n f ( x ) = i = 1 D | x i | ( i + 1 ) 1 x i 1 0
07ChungReynolds M i n f ( x ) = ( i = 1 D x i 2 ) 2 100 x i 100 0
08Csendes M i n f ( x ) = i = 1 D x i 6 ( 2 + sin 1 x i ) 1 x i 1 0
09Schaffer M i n f ( x ) = 0.5 + sin 2 ( x 1 2 + x 2 2 ) 2 0.5 1 + 0.001 ( x 1 2 + x 2 2 ) 2 100 x i 100 0
10Schumer_Steiglitz M i n f ( x ) = i = 1 D x i 4 100     xi     100 0
11Schwefel M i n f ( x ) = ( i = 1 D x i 2 ) α 100 x i 100 0
12Schwefel1.2 M i n f ( x ) = i = 1 D ( j = 1 i x j ) 2 100 x i 100 0
13Schwefel 2.21 M i n f ( x ) = max 1 i D | x i | 100 x i 100 0
14Schwefel 2.22 M i n f ( x ) = i = 1 D | x i | + i = 1 D | x i | 100 x i 100 0
15Schwefel 2.23 M i n f ( x ) = i = 1 D x i 10 10 x i 10 0
16Zakharov M i n f ( x ) = i = 1 D x i 2 + ( 1 2 i = 1 n i x i ) 2 + ( 1 2 i = 1 n i x i ) 4 5   x i   10 0
D shows the dimensionality of the problem, S represents the interval of the variables, and fmin denotes the standard global optimum minimum value. In the parameters for the simulation used as c1 = c2 = 1.45, inertia weight w is used in the interval [0.9, 0.4], and swarm size is 50. For simulation, the function dimensions are D = 10, 20, and 30, and the maximum number of epochs is 3000.
Table 4. Comparative results for all PSO-based approaches on 16 standard benchmark functions.
Table 4. Comparative results for all PSO-based approaches on 16 standard benchmark functions.
FunctionsDIM × ItrPSOSO-PSOH-PSOTO-PSOWE-PSOKN-PSO
MeanMeanMeanMeanMeanMean
F110 × 10002.33 × 10−742.74 × 10−763.10 × 10−775.57 × 10−785.91 × 10−780.0000 × 10+00
20 × 20001.02 × 10−848.20 × 10−881.76 × 10−901.30 × 10−904.95 × 10−903.14001 × 10−217
30 × 30001.77 × 10−267.67 × 10−204.13 × 10−321.25 × 10−511.30 × 10−428.91595 × 10−88
F210 × 10004.97 × 10−014.97 × 10−017.96 × 10−013.98 × 10−012.98 × 10−01−8602.02
20 × 20008.17 × 10+006.47 × 10+003.58 × 10+002.89 × 10+003.11 × 10+00−31,433.3
30 × 30001.01 × 10+019.86 × 10+009.45 × 10+008.16 × 10+007.76 × 10+00−60,711.8
F310 × 10008.70 × 10−801.79 × 10−794.87 × 10−793.91 × 10−824.40 × 10−810.0000 × 10+00
20 × 20002.621447.864322.621447.07 × 10−901.78 × 10−894.78718 × 10−237
30 × 30002.62 × 10+011.57 × 10+011.05 × 10+017.70 × 10−353.87 × 10−571.57084 × 10−97
F410 × 10004.46 × 10−1473.86 × 10−1479.78 × 10−1457.29 × 10−1481.24 × 10−1500.0000 × 10+00
20 × 20003.14 × 10−1559.27 × 10−1542.75 × 10−1595.14 × 10−1584.96 × 10−1590.0000 × 10+00
30 × 30001.82 × 10−1332.36 × 10−1358.53 × 10−1303.13 × 10−1382.54 × 10−1361.6439 × 10−228
F510 × 10004.35 × 10−798.95 × 10−792.43 × 10−782.04 × 10−802.20 × 10−800.0000 × 10+00
20 × 20001.31 × 10+013.93 × 10+011.31 × 10+013.54 × 10−893.12 × 10−892.39359 × 10−236
30 × 30001.31 × 10+027.86 × 10+015.24 × 10+013.85 × 10−341.94 × 10−562.9093 × 10−87
F610 × 10001.70 × 10−614.45 × 10−647.29 × 10−662.46 × 10−664.62 × 10−663.04226 × 10−318
20 × 20003.25 × 10−1124.39 × 10−1125.01 × 10−1092.56 × 10−1154.45 × 10−1138.59557 × 10−277
30 × 30007.21 × 10−1354.10 × 10−1241.51 × 10−1346.22 × 10−1376.96 × 10−1352.33033 × 10−223
F710 × 10002.96 × 10−1572.39 × 10−1571.28 × 10−1574.89 × 10−1592.47 × 10−1630.0000 × 10+00
20 × 20008.79 × 10−1771.77 × 10−1843.49 × 10−1833.09 × 10−1873.41 × 10−1860.0000 × 10+00
30 × 30001.23 × 10−821.25 × 10−1165.99 × 10−1305.01 × 10−1354.60 × 10−1348.03288 × 10−175
F810 × 10004.39 × 10−2001.98 × 10−1944.51 × 10−1971.26 × 10−2028.99 × 10−2014.9228 × 10−67
20 × 20001.57 × 10−201.04 × 10−931.10 × 10−1482.84 × 10−1574.09 × 10−1514.5887 × 10−16
30 × 30001.89 × 10−094.54 × 10−101.14 × 10−081.40 × 10−101.34 × 10−092.2334 × 10−08
F910 × 10005.49 × 10−011.30 × 10−012.02 × 10−011.26 × 10−011.42 × 10−010.824968
20 × 20002.05 × 10+007.83 × 10−016.83 × 10−015.84 × 10−014.32 × 10−014.56265
30 × 30001.12 × 10+009.99 × 10−019.56 × 10−019.06 × 10−019.12 × 10−017.25675
F1010 × 10002.23 × 10−1382.23 × 10−1384.35 × 10−1371.02 × 10−1401.10 × 10−1390.0000 × 10+00
20 × 20003.79 × 10−1487.87 × 10−1494.19 × 10−1473.78 × 10−1518.73 × 10−1530.0000 × 10+00
30 × 30004.43 × 10−1267.52 × 10−1331.57 × 10−1282.03 × 10−1341.38 × 10−1332.26229 × 10−221
F1110 × 10003.75 × 10−1871.57 × 10−1922.15 × 10−1915.57 × 10−1988.99 × 10−1980.0000 × 10+00
20 × 20005.29 × 10−1932.53 × 10−1958.45 × 10−1958.45 × 10−1959.83 × 10−1970.0000 × 10+00
30 × 30004.82 × 10−1548.84 × 10−1595.49 × 10−1682.04 × 10−1705.75 × 10−1739.00586 × 10−278
F1210 × 10001.13 × 10−011.67 × 10−022.28 × 10−024.78 × 10−032.89 × 10−032.739 × 10−12
20 × 20001.39 × 10+015.03 × 10+002.95 × 10+001.28 × 10+001.67 × 10+007.819 × 10+00
30 × 30007.45 × 10+001.22 × 10+018.74 × 10+002.94 × 10+004.94 × 10+002.239 × 10+01
F1310 × 10008.04 × 10−268.01 × 10−273.59 × 10−271.24 × 10−271.41 × 10−270.0000 × 10+00
20 × 20001.42 × 10−082.64 × 10−113.29 × 10−102.99 × 10−102.14 × 10−120.0000 × 10+00
30 × 30006.20 × 10−031.41 × 10−039.36 × 10−031.12 × 10−031.41 × 10−030.0000 × 10+00
F1410 × 10003.62 × 10−383.62 × 10−385.92 × 10−366.92 × 10−391.95 × 10−387.78286 × 10−197
20 × 20006.27 × 10−101.38 × 10−097.91 × 10−132.49 × 10−121.17 × 10−136.6163 × 10−12
30 × 30002.56 × 10−064.80 × 10+011.34 × 10−065.40 × 10−114.88 × 10−099.3032 × 10−06
F1510 × 10001.10 × 10−2943.19 × 10−3012.78 × 10−3071.94 × 10−3073.21 × 10−3086.26612 × 10−138
20 × 20006.16 × 10−2715.09 × 10−2763.74 × 10−2701.60 × 10−2764.85 × 10−2681.29033 × 10−25
30 × 30003.08 × 10−2071.04 × 10−2008.12 × 10−2092.34 × 10−2153.06 × 10−2122.27 × 10−06
F1610 × 10005.48353858.5299 × 10−173.3074 × 10−161.2248038.3354 × 10−072.26476 × 10−27
20 × 200083.4671.63440.1803749.168415.13227.17014 × 10−72
30 × 3000265.90708282.186445.0408133.967967.03015.45179 × 10−251
Table 5. Mean ranks obtained by Kruskal–Wallis and Friedman test for all PSO-based approaches.
Table 5. Mean ranks obtained by Kruskal–Wallis and Friedman test for all PSO-based approaches.
Friedman ValueKruskal–Wallis
PSO39.0939.33
SO-PSO37.4738.39
H-PSO38.5038.91
TO-PSO41.7942.67
WE-PSO41.8842.50
KN-PSO18.2423.31
Table 8. Characteristics of UCI benchmarks data set.
Table 8. Characteristics of UCI benchmarks data set.
Features
Sr. NoData SetContinuousNatureNo. of InputsNo. of Classes
1Diabetes8Real82
2Heart13Real132
3Wine13Real133
4Seed7Real73
5Vertebral6Real62
6Blood Tissue5Real52
7Mammography6Real62
Table 9. Results of 10-fold classification rates of ANN-training methods in for seven data sets for accuracy.
Table 9. Results of 10-fold classification rates of ANN-training methods in for seven data sets for accuracy.
Sr. NoData SetsTypeBPAPSONNSO-PSONNH-PSONNTO-PSONNWE-PSONNKN-PSONN
Ts. AccTs. AccTs. AccTs. AccTs. AccTs. AccTs. Acc
1Diabetes2-Class65.3%69.1%69.1%71.6%73.3%74.1%78.5%
2Heart2-Class68.3%72.5%67.5%72.5%77.577.5%79%
3Wine3-Class62.17%61.11%66.66%67.44%69.44%69.6%72%
4Seed3-Class70.56%77.77%84.44%77.77%88.88%91.11%93%
5Vertebral2-Class84.95%92.85%92.85%92.85%94.64%94.64%96%
6Blood Tissue2-Class73.47%78.6%78.66%70%82.66%84%87%
7Memo Graphy2-Class71.26%76.66%63%85%88.88%96.66%98%
Results of 10-fold classification rates of ANN-training methods in for seven data sets for accuracy.
Table 10. One way ANOVA Test Results of PSO variants.
Table 10. One way ANOVA Test Results of PSO variants.
ParameterRelationSum of SquaresdfMean SquareFSignificance
Testing AccuracyAmong groups1318.26219.6972.36760.04639
Table 11. Results of 10-fold classification rates of ANN-training methods in for seven data sets for accuracy.
Table 11. Results of 10-fold classification rates of ANN-training methods in for seven data sets for accuracy.
Sr. NoData SetsTypeBPADEDE-SDE-WEDE-TODE-KNDE-H
Ts. AccTs. AccTs. AccTs. AccTs. AccTs. AccTs. Acc
1Diabetes2-Class65.3%66.1%68.16%69.6%71.30%67.17%75.50%
2Heart2-Class68.3%70.5%72.5%71.5%74.50%72.56%76.34%
3Wine3-Class62.17%64.7%65.19%66.20%66.59%68.25%70.51%
4Seed3-Class70.56%75.16%75.29%75.77%82.13%86.76%91.54%
5Vertebral2-Class79.95%82.13%84.26%86.15%87.64%90.17%96.25%
6Blood Tissue2-Class73.47%76.23%74.16%72..21%84.76%81.34%86.45%
7Mammography2-Class71.26%74.39%68.37%82.45%86.17%96.66%99.21%
Results of 10-fold classification rates of ANN-training methods in for seven data sets for accuracy.
Table 12. One way ANOVA test results of DE variants.
Table 12. One way ANOVA test results of DE variants.
ParameterRelationSum of SquaresdfMean SquareFSignificance
Testing AccuracyAmong groups1180.06196.6722.84530.02043
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Bangyal, W.H.; Nisar, K.; Ag. Ibrahim, A.A.B.; Haque, M.R.; Rodrigues, J.J.P.C.; Rawat, D.B. Comparative Analysis of Low Discrepancy Sequence-Based Initialization Approaches Using Population-Based Algorithms for Solving the Global Optimization Problems. Appl. Sci. 2021, 11, 7591. https://0-doi-org.brum.beds.ac.uk/10.3390/app11167591

AMA Style

Bangyal WH, Nisar K, Ag. Ibrahim AAB, Haque MR, Rodrigues JJPC, Rawat DB. Comparative Analysis of Low Discrepancy Sequence-Based Initialization Approaches Using Population-Based Algorithms for Solving the Global Optimization Problems. Applied Sciences. 2021; 11(16):7591. https://0-doi-org.brum.beds.ac.uk/10.3390/app11167591

Chicago/Turabian Style

Bangyal, Waqas Haider, Kashif Nisar, Ag. Asri Bin Ag. Ibrahim, Muhammad Reazul Haque, Joel J. P. C. Rodrigues, and Danda B. Rawat. 2021. "Comparative Analysis of Low Discrepancy Sequence-Based Initialization Approaches Using Population-Based Algorithms for Solving the Global Optimization Problems" Applied Sciences 11, no. 16: 7591. https://0-doi-org.brum.beds.ac.uk/10.3390/app11167591

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop