Next Article in Journal
The Third and Fourth Kind Pseudo-Chebyshev Polynomials of Half-Integer Degree
Next Article in Special Issue
Kernel Ridge Regression Model Based on Beta-Noise and Its Application in Short-Term Wind Speed Forecasting
Previous Article in Journal
Soliton and Breather Splitting on Star Graphs from Tricrystal Josephson Junctions
Previous Article in Special Issue
Tenant-Oriented Monitoring for Customized Security Services in the Cloud
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Data Source Selection Based on an Improved Greedy Genetic Algorithm

1
School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
2
Research Institute of Information, Beijing National Research Center for Information Science and Technology, Department of Computer Science and Technology, Institute of Internet Industry, Tsinghua University, Beijing 100084, China
*
Author to whom correspondence should be addressed.
Submission received: 16 January 2019 / Revised: 18 February 2019 / Accepted: 18 February 2019 / Published: 20 February 2019
(This article belongs to the Special Issue Information Technology and Its Applications 2021)

Abstract

:
The development of information technology has led to a sharp increase in data volume. The tremendous amount of data has become a strategic capital that allows businesses to derive superior market intelligence or improve existing operations. People expect to consolidate and utilize data as much as possible. However, too much data will bring huge integration cost, such as the cost of purchasing and cleaning. Therefore, under the context of limited resources, obtaining more data integration value is our expectation. In addition, the uneven quality of data sources make the multi-source selection task more difficult, and low-quality data sources can seriously affect integration results without the desired quality gain. In this paper, we have studied how to balance data gain and cost in the source selection, specifically, maximizing the gain of data on the premise of a given budget. We proposed an improved greedy genetic algorithm (IGGA) to solve the problem of source selection, and carried out a wide range of experimental evaluations on the real and synthetic dataset. The empirical results show considerable performance in favor of the proposed algorithm in terms of solution quality.

1. Introduction

With the continuous development of information technology, data is always being produced in all fields of modern society at all times, especially in some industries with huge data volume, such as telecommunications, transportation, medical care, securities and so on, which will generate a huge amount of data in a short period of time. To make full use of data resources and improve competitiveness in the field, policymakers need to integrate data resources and increase sources of information for deeper data analysis and mining efforts, which can bring huge benefits. However, there is no such thing as a free lunch. Data collection and integration will be costly before they bring us profit, for instance, currently emerging data platforms include Factual [1], Infochimps [2], Xignite [3], Windows Azure Data Marketplace [4], etc., which require to be paid. Even for a free and open dataset, it takes a lot of time and energy to clean up the data and solve the problem of heterogeneous data conflicts. The costs are high for individuals or companies. Therefore, in the data integration process, it is more common practice to collect data sources with high data quality and wide coverage as much as possible without exceeding the limited budget. Therefore, how to balance data integration cost and maximize integrated revenue are important issues. The existing literatures [5,6,7] have proved that the data source selection is a NP hard problem, which can be summarized as a 0–1 knapsack problem. Currently, various exact algorithms have been proposed to solve the 0–1 knapsack problem, including dynamic programming [8], core algorithms, branch and bound [9], etc. Since the solution space size of the problem is exponential with the problem input scale, these exact algorithms are not suitable for solving the larger 0–1 knapsack problem. Therefore, for the data source selection problem, a special heuristic solution is especially needed to solve the problem of selecting a suitable target source from a large amount of information sources. Heuristics are known to be classified into the following categories; binary particle swarm optimization(BPSO) [10], ant colony algorithm (ACA) [11], genetic algorithms [12], etc.
Along this line of thought, we formalized the problem of source selection as a 0–1 knapsack problem [13] and chose the appropriate solution in all possible combinations. We proposed a gain-cost model driven by data quality and used intelligent approaches to deal with this complex problem, especially with genetic algorithms (GAs). GAs, which are considered robust and efficient [14], have been widely used and are better than the methods of large-scale data analysis [15]. However, the global optimization ability and execution efficiency of genetic algorithms are often not ideal. Literature [16] believes that the reason why the genetic algorithm is not efficient to resample the points visited in the search space, is essentially caused by the randomness of genetic operators (selection, crossover, mutation). In this paper, we make appropriate improvements for these operators, and propose a novel greedy strategy to solve the source selection problem, which makes the performance of the genetic algorithm more efficient.
In short, the key contributions of this paper can be summarized as follows:
  • We first summarize several dimensional indicators that affect the quality of data and establish a linear model to estimate the quality score. A gain-cost model driven by integration scores was proposed, which provides the basis for assessing the value of data sources.
  • We propose an improved novel greedy genetic algorithm(IGGA). Not only improved genetic operators, but also a novel greedy strategy are proposed, which makes the source selection problem more efficient.
  • We have conducted extensive experiments on real and synthetic datasets. A large number of experimental results show that our algorithm is very competitive against the other state-of-the-art intelligent algorithms in terms of performance and problem solving quality.
The remainder of this study is organized as follows. Related source selection methods and genetic algorithm in combinatorial optimization work is covered in Section 2. In Section 3, we present the problem formulation and model of the data source selection, propose the methods of data quality and source coverage estimation, and establish a gain-cost model driven by a comprehensive score. We improve the genetic operator and design an efficient greedy strategy for source selection in Section 4. We evaluate the performance of our algorithms through real and synthetic datasets in Section 5. Section 6 summarizes and discusses future research directions.

2. Related Works

2.1. The Sources Selection Approaches in a Distributed Environment

For the issue of data source selection, some classic methods have been developed. There is a wealth of literature on this topic, and then we present some of the relevant results in this area.
Much work has been done for online data consolidation, especially for source selection of deep network [17,18], but most of the work is just focused on finding the data source for a given query or domain. Such work can be summarized as a document retrieval method. Specifically, the data source is represented as a file connection or sample a document for indexing, and according to the information retrieval technology, the returned documents are classified and sorted according to the similarity of the query keywords. Then the information source is selected. Related researches include [19,20,21,22]. In recent years, people also tried to design intelligent algorithms in the field of information retrieval. Genetic algorithms are widely used to modify document descriptions, user queries, and adapt matching functions. For example [15,23,24]. However, most of these studies did not consider the impact of data source quality on the result of the source selection. Much work [25,26] focused on turning data quality standards into optimization goals for query decisions in every situation and using this information to improve the quality of query results in the data integration process. However, none of them studied the effect of source cost on the selection results. Compared with online data integration, offline data integration is less studied. Dong et al. [5] focused on the marginal revenue standard of data sources to balance data quality and integration cost. In the source selection process, the focus is to select a subset of sources for data integration, so that the overall profit of the selected source is the highest. Although they had done a great number of experimental researches, no further discussion was conducted on large sample datasets.

2.2. Application of a Genetic Algorithm in Combinatorial Optimization

As the size of the problem increases, the search space for combinatorial optimization problem also expands dramatically. Sometimes it is difficult or even impossible to find the exact optimal solution using the enumeration method on current computers. For such complex issues, people have realized that their main energy should focus on seeking the satisfied solution, and genetic algorithm is one of the best tools to find this solution. Practice has proved that the genetic algorithm has been successfully applied in solving source selection problem. Lebib et al. [27] proposed a method based on a genetic algorithm and social tagging to select, with the optimal possible way, data sources to be interrogated. Kumar et al. [28] use a genetic algorithm to select the appropriate search engine for the user query in the meta search engine. Since the user’s information needs are stored in the database of different underlying search engines, the choice of the search engine substantially improves the user’s query efficiency. Abououf et al. [29] address the problem of multi-worker multi-task selection allocation for mobile crowdsourcing, and use genetic algorithms to select the right workers for each task group, looking to maximize the QoS of the task while minimizing the distance traveled.
In addition, genetic algorithm has been applied to solve various NP-hard problems, such as the traveling salesman problem, knapsack problem, packing problem, etc. LarrNaga et al. [30] developed a genetic algorithm with different representations to tackle the travelling salesman problem. Lim et al. [31] borrowed from social monogamy: pair bonding and infidelity at a low probability and explored a pair of genetic algorithms to solve the 0–1 knapsack problem, and achieved better results. Quiroz-Castellanos et al. [32] proposed a method that was referred to the grouping genetic algorithm with controlled gene transmission to solve the bin packing problem.
To our knowledge, very few works address the problem of offline sources selection. The work of [5] is close to our work.

3. Data Source Selection Driven by the Gain-Cost Model

3.1. Problem Definition

Before defining the problem, we need to make some assumptions. We considered integrating from a set of data source S, assuming that the data integration system provided the functions of measuring cost and gain. For the cost, on the one hand, it is related to the cost of purchasing data from a specific source, depending on the pricing mechanism of the data platform. On the other hand, cost is related to data cleansing, or any other foreseen expense in the process of data integration. For such costs, historical data can be used for estimation, and the cost incurred in the data integration process can be obtained. The gain consisted of two factors, one was determined by the quality of the data source, such as the completeness and accuracy of the data, and the other was determined by the coverage [33], i.e., the data source containing the number of entities.
Before we proceed any further, it will be helpful to define a few terms. We first define some features of data source formally. Let g i , c i , Q i , C o v i be the gain, cost, quality score and coverage of i-th data source. Q i , C o v i and g i are related, and we will give a detailed explanation in the next subsection. Considering the above factors, we use g i to represent the comprehensive gain of i-th source. Then a set of data source s is given, G ( s ) = i = 1 | s | g i , C ( s ) = i = 1 | s | c i . Next we define the problem as follows.
Definition 1.
(Source selection) Let Ω be a set of sources, Ω = s 1 , s 2 , , s n , and δ c be a budget on cost. The source selection problem finds a subset S ⊆ Ω that maximizes G ( S ) under constraint C ( S ) δ c , which can be described as follows:
M a x i m i z e G ( s ) = i = 1 | s | g i = i = 1 n g i x i s . t . = C ( s ) = i = 1 | s | c i = i = 1 n c i x i δ c x i = 0 o r 1 , i = 1 , 2 , , n | s | n .
The binary decision variable x i is used to indicate whether data source i is selected. It may be assumed that all gain and cost are positive, and that all cost is smaller than the budget δ c .

3.2. Data Quality and Coverage

In this paper, we considered data quality from multiple aspects. Specifically, we evaluateed quality in three areas, i.e., completeness, redundancy, accuracy, which are denoted by A, B and C respectively. Table 1 [34] contains the metrics we defined for each of the selected quality attributes, reporting names and descriptions and lists the formulas used to calculate them.
Then, we made a weighted average of these three attributes as follows:
Q ( s i ) = α A + β B + γ C s . t . α + β + γ = 1 ,
where Q ( s i ) is the quality score of the data source s i , α , β , γ are the weight of each attribute, which can be set by the user.
Definition 2.
(Coverage) Let s i be the selected set of sources to be integrated and count the number of s i containing entities as | s i | . We define the coverage of s i , denoted by C o v ( s i ) , as the probability that a random entity from the world Ω. We express this probability as:
C o v ( s i ) = | s i | Ω .
Example 1.
We consider data sets obtained from online bookstores. We hope to collect data on computer science books. At present, there are 894 bookstores offering a total of 1265 computer science books (each bookstore corresponds to one data provider). We pay attention to coverage, i.e., the number of books provided. After inspection, the largest data source provides 1096 books, so the coverage of the data source can be calculated to be 1096/1265 = 0.86.
Next we make some discussions for the comprehensive score of i-th source, denoted by I ( s i ) . We analyze both data quality and coverage, and assume that they are independent of each other. The quality of data depends on the completeness, redundancy and accuracy. The coverage of data source s i is expressed in the data source containing the number of entities. On the one hand, high coverage with near zero quality should have a very low comprehensive score, and on the other hand, very high quality with near zero coverage should also have a very low comprehensive score. Therefore, its comprehensive score will be high only if the quality and coverage are both high, which is consistent with our intuition. Based on the discussions above, the comprehensive score I ( s i ) of i-th source is written as follows:
I ( s i ) = Q ( s i ) × C o v ( s i )

3.3. Gain-Cost Models

We consider the impact of different gain-cost models on source selection, and adopt the comprehensive score of Section 3.2. Treat quality score and coverage as important gain factors, and establish two gain models.
  • Linear gain assumes that the gain grows linearly with a certain composite score metric I ( s ) and sets g ( s ) = 100 I ( s )
  • Step gain assumes that reaching a milestone of quality will significantly increase the gain and set:
    g ( s ) = 100 I ( s ) : 0 I ( s ) < 0 . 2 100 + 100 ( I ( s ) - 0 . 2 ) : 0 . 2 I ( s ) < 0 . 5 150 + 100 ( I ( s ) - 0 . 5 ) : 0 . 5 I ( s ) < 0 . 8 200 + 100 ( I ( s ) - 0 . 8 ) : 0 . 8 I ( s ) < 1
We assign the cost of a source in [5, 20] in two ways:
  • Linear costassumes the cost grows lineraly with the I ( s ) and applies c ( s ) = 15 I ( s ) + 5 ;
  • Step cost assumes reaching some milestone of I ( s ) will significantly increase cost and so applies:
    c ( s ) = 5 : 0 I ( s ) < 0.2 10 : 0.2 I ( s ) < 0.5 15 : 0.5 I ( s ) < 0.8 20 : 0.8 I ( s ) 1
By randomly combining the above gain and cost models, we obtain the following four gain and cost generation models, summarized in the following Table 2.
Meanwhile, we set the budget δ c = λ i = 1 | s | c ( s i ) , and the λ is a random real number on [ 0.3 , 0.6 ] . In the experiment, we set λ as 0.5.
Then we use an example to illustrate the calculational method of data source gain and cost.
Example 2.
Table 3 shows the employee information provided by one data source. First, we calculate the data quality score. According to the calculation method of data quality in Table 1, we can conclude that A is (40 − 4)/40 = 0.9, where the empty element value is four and the redundancy B is 1/5 = 0.2. This is because when the two elements of ID1 and ID9 conflict with each other, we only choose either of them. Since this factor is the cost-indicator, we convert it to the benefit-indicator, i.e., 1 − 0.2 = 0.8. Under the knowledge that the regional codes of Beijing, Shanghai and Guangzhou are 010, 021 and 020, respectively, ID3 and ID6 violate the rules. According to the formula of accuracy, C = (40 − 2)/40 = 0.95. For simplicity, we set three coefficients α, β and γ to 0.3, 0.3 and 0.4, respectively. Thus Q(s) = 0.3A + 0.3B + 0.4C = 0.9 × 0.3 + 0.8 × 0.3 + 0.95 × 0.4 = 0.89. Secondly, we calculate the comprehensive score. Here we only describe one data source. In fact, there will be multiple data sources providing employee information, assuming that all sources can provide a maximum of 10 entities, while Table 3 only provides five entities. The coverage based on Equation (3) can be calculated as 5/10 = 0.5. Thus, the comprehensive score I ( s i ) = Q ( s i ) × C o v ( s i ) = 0 . 89 × 0 . 5 = 0 . 445 . Finally, calculate the gain and cost of the data source based on the model provided in Section 3.3. Taking the linear gain-cost model as an example, thus g(s) = 100 × 0.445 = 44.5, c(s) = 15 × 0.445 + 5 = 11.675.

4. Improved Greedy Genetic Algorithm (IGGA)

The paper makes improvement for the genetic algorithm in the following three aspects:

4.1. Change the Way of Selection

Genetic algorithms use a variety of selection functions, including level selection, steady state selection, elite retention, and roulette-wheel selection. Some studies have shown that the roulette method is ideal for implementing selection operators and this approach enhances the chances of being suitable for chromosome selection. All solutions are placed on the roulette wheel, and a better solution has a larger portion on the roulette wheel, which provides a fair chance for each solution, and the probability of being selected is proportional to the fitness value, so individuals with higher fitness have a higher probability of survival, a higher chance of being chosen. Assuming that M i ( i = 1 , 2 , , s ) is a chromosome in a population, z is the population size, and f ( M i ) is the fitness of M i . Fitness is expressed as the gain of the data source in Section 3.3. The selection probability ( P r o b ) of each chromosome is calculated by Equation (5). The sum of probabilities from 1 to i is denoted by Equation (6).
P r o b i = f ( M i ) / i = 1 z f ( M i )
S u m i = j = 1 i P r o b j
Algorithm 1 describes our selection process. In order to preserve the best chromosomes, we improved the selection algorithm. All individuals in the population were ranked in descending order of fitness, with the top 1/4 of the individuals being replicated twice, the middle 2/4 individual being kept, and discarding the last 1/4. In this way, on the one hand, individuals with lower fitness can be directly eliminated, and on the other hand, the proportion of individuals with better fitness can be increased.
Algorithm 1 Selection.
input: All members of population
output: New Selected population
1 Sort all data sources in descending order of fitness;
2 The individual in the top 1/4 is copied twice, the middle 2/4 is kept, the last 1/4 is abandoned and generate a transition population;
3 Generate a random number R [0,1];
4 repeat
Symmetry 11 00273 i001
8 until create offspring;

4.2. Crossover

Crossover functions were used for generation of new chromosomes. By recombining and distributing the genes on the parental chromosome to generate the children, crossover may bring together the parents’ dominant genes to produce new individuals that are more adaptive and closer to the optimal solution. We used partial-mapped crossover without “duplicates”. Thus, the gene in the generated chromosome must not be repeated. Details are given in Algorithm 2.
Algorithm 2 Crossover.
input: Parents from the current population.
output: Two new children.
1 Let M = ( m 1 , m 2 , m k ) and N = ( n 1 , n 2 n k ) two parents to crossed;
2 Choose two random number(a,ba<b) on the set {1,2, …k}, two new children M ¯ and N ¯ are created according to the following rules:
M ¯ = n i i f ( a < i < b ) m i o t h e r w i s e N ¯ = m i i f ( a < i < b ) n i o t h e r w i s e
3 Establishing the mapping relationship in segment (a, b), f(M) ↔f(N);
4 Remove, before the cutting point(a) and after the cutting point(b), the data source which are already placed segment (a,b);
5 Put the corresponding data source on the delete location according to the mapping.
Example 3.
We consider 10 data sources, a flag of 1 indicates that the source was selected, a flag of 0 indicates that the source was not selected. Figure 1 shows an example of crossover.
  • Step 1. Randomly select the starting and ending positions of several genes in a pair of chromosomes (the two chromosomes are selected for the same position).
  • Step 2. Exchange the location of these two sets of genes.
  • Step 3. Detect conflict, according to the exchange of two sets of genes to establish a mapping relationship. Taking S 3 - S 7 - S 1 as an example, we can see that there are two genes S 3 in proto-child two in the second step, when it is transformed into the gene S 1 by the mapping relationship, and so on. Finally, all the conflicting genes are eliminated to ensure that the formation of a new pair of offspring genes without conflict.
  • Step 4. Finally get the result.

4.3. Novel Greedy Repair Strategy

In the randomly generated initial population and each generation of genetically derived populations, there will always be some individuals who do not meet the constraints, i.e., the data source cost corresponding to their chromosomes exceeds the budget limit. A novel greedy strategy is proposed, which is named the greedy repair strategy (GRS) and given in Algorithm 3. Before the start of the algorithm, all data sources were sorted in descending order according to the ratio of comprehensive gain and cost, and the subscripts of each item were stored in array A[0⋯n] according to the sorted order. Let F l a g [ i ] be a boolean array that identifies the state of each data source. When F l a g [ i ] = 1, the data source was selected and when F l a g [ i ] = 0, it was not selected. The Algorithm 3 was first introduced in descending order of gain and cost and stored in A [ y i ] , then we selected data sources in turn, changed F l a g [ i ] = 1, and calculated the cumulative gain and cost. It is worth mentioning that our algorithm differed from other traditional greedy algorithms in that when the cumulative cost was greater than our pre-set budget, our algorithm did not stop. Instead, the cost of the currently selected data source was subtracted from the total cost and the identification of the data source was changed to 0. In this way, the previous steps were repeated again if the cumulative cost was less than the budget until all data sources were detected. The final outputs were new chromosomes and total gain, denoted by Y and G a i n ( Y ) , respectively.
Example 4.
To illustrate, we list six data sources to be selected in Table 4 and arranged them in non-ascending order of the gain-cost ratio, assuming that the budget was 100. According to the traditional greedy strategy, after selecting the source s 3 , the algorithm will stop executing. This is because if the algorithm continues to select s 4 , the total cost will exceed the budget. Therefore, the traditional greedy algorithm gets the total gain is (90 + 80 + 75 = 245). When using GRS, our algorithm will not stop when it executes to source s 4 . On the contrary, GRS will skip s 4 and continue to evaluate source s 5 . GRS will continue to execute until the total cost does not exceed the budget. Search for a complete list of alternate data sources. At this time, the result obtained by GRS is (90 + 80 + 75 + 45 + 10 = 300). It can been seen that GRS has an advantage over traditional greedy strategies.
Algorithm 3 Greedy repair strategy (GRS).
input: Chromosome S = [ s 1 , s 2 ,…, s n ], A[0…n], δ c : cost budget
output: A new chromosome Y = [ y 1 , y 2 ,…, y n ], G a i n ( Y )
1 y i 0, F l a g [ i ] 0;
2 G a i n 0, C o s t 0, i 0;
3 Arrange data sources in descending order of gain-cost ratio;
4 for ( i 1 to n) do
Symmetry 11 00273 i002
11 return (Y, G a i n ( Y ) )

4.4. Integrate Greedy Strategy into GAs

Based on the above analysis, it can been seen that if GRS is used to optimize the data source selection process, not only can the problem be solved, but also the entire data source can be traversed to make the optimal solution. Next, we integrate GRS into the improved genetic algorithm, which is named IGGA. The detailed pseudocode is given in Algorithm 4.
Algorithm 4 Improved greedy genetic algorithm (IGGA).
input S = { s 1 , s 2 , , s n } ,
    G = {( g 1 , g 2 , g i , , g n )|( 0≤ i ≤ n)},
    C = {( c 1 , c 2 , c i , , c n )| (0≤ i ≤ n)},
    Z, p c , p m , M a x I t
output: Optimal solution S ( r ) , Objective gain value f ( S ( r ) )
1 A[0…n]←Descending{ g i / c i | g i G, c i C, (0 ≤ i ≤ n)};
2 Generate randomly an initial population of P = { S i (0) | 1 ≤ i ≤ Z};
3 for i← 1 to Z do
Symmetry 11 00273 i003
5 Set the number of inner loops r ← 0;
6 while (r ≤ M a x I t ) do
Symmetry 11 00273 i004
14 return S ( r ) , f ( S ( r ) ) ;
The genetic algorithm based on the above greedy strategy is as follows:
  • Step 1. According to the greedy repair strategy, all the selected data sources are non-incrementally sorted according to the ratio of gain and cost.
  • Step 2. Use the binary coding method, randomly generate the initial population P, and use the greedy strategy to obtain the initial current optimal solution.
  • Step 3. The fitness is calculated for each chromosome in the population P, and if the value corresponding to the chromosome is greater than the current optimal solution, the current solution is replaced.
  • Step 4. If the maximum number of iterations is reached, then stop. Otherwise, a crossover operation is performed, and a temporary population P 1 ( r ) is obtained according to the crossover probability.
  • Step 5. With a small probability p m , a certain gene of each chromosome is mutated, and then a temporary population P 2 ( r ) is generated. Use greedy strategies to repair chromosomes that do not meet the constraints.
  • Step 6. Select some chromosomes according to Equation (5) to form a new population P ( r + 1 ) , and turn to Step 3.
In our algorithm, genetic operations (i.e., selection, crossover, mutation) further explored and utilized more combinations to optimize the objective function, while the greedy repair strategy not only improved the efficiency of the algorithm, but also evaluated whether each candidate data source met the constraints, so as to obtain high-quality solutions.

5. Experimental Design and Result

5.1. Experimental Design

This section will include algorithm parameter settings, and a number of comparative and verification experiment listed below:
  • To make the comparison as fair as possible, we discussed the trend of IGGA parameter values under the four gain-cost models and set reasonable values for them.
  • We compared the average performance and the stability of IGGA, DGGA [35], BPSO [10] and ACA [11] in a real data set.
  • We compared the performance of IGGA and DGGA in the synthetic data set. In addition, we made comparisons with other state-of-the-art intelligent algorithms to verify the efficiency of IGGA.
All experiments were coded in Python under Windows 10 for Education platform on an Intel Core i7 2.8 GHz processor with 8 GB of RAM.

5.1.1. Dataset

We employed both real and synthetic datasets in the experimental evaluations.
We experimented on two data sets, i.e., Book and Flight [36]. The Book contained 894 data sources, which were registered with AbeBooks.com and in 2007 provided information on computer science books. They provided a total of 12,436 books, with ISBNs, names, and authors. The data source coverage provided was from 0.1% to 86%. The Flight collected information of over 1200 flights from 38 sources over a one-month period, together providing more than 27,000 records, each source providing 1.6% to 100% of the flights.
In addition, we used the classic datasets of the 0–1 knapsack problem, which were provided by [37,38]. The size of these datasets was less than 40 problem instances. Here, we called these small-S. Meanwhile, according to the method provided by [37], we also randomly generated eight instance sets ranging from 100 to 1500, which were called large-S.

5.1.2. Parameter Settings

In the experiment, two data sets were applied to four gain-cost models and eight data instances were generated. Since the IGGA is a kind of parameter-sensitive evolutionary algorithm, crossover operation and mutation operation played key roles in the generation of new solutions. Many researches [39] have shown that it is difficult to search forward when the crossover probability p c is too small, and it is easy to destroy the high fitness structure. When the mutation probability p m is too low, it is difficult to generate a new gene structure. Conversely, if p m is too high, GA becomes a simple random search. To further investigate the optimal values of p c and p m in IGGA, we solved eight instances by IGGA to determine the values of p c and p m .
Specifically, we set p c to (0.1, 0.3, 0.5, 0.7, 0.9), and p m to (0.01, 0.05, 0.1, 0.3, 0.5, 0.7, 0.9), respectively. As a result, we got a total of 35 different combinations ( p c , p m ), and an ID was assigned to each combination, which are shown in Table 5. Once the range of parameters was determined, the different combinations of parameters on each gain-cost model was independently calculated 30 times. Finally, we got a performance comparison of IGGA with 35 parameters on the four gain-cost models (in Section 4). According to Figure 2 and Figure 3, it is easy to see that p c = 0.5 and p m = 0.01 are the most reasonable choices.
When eight instances are solved with IGGA, DGGA, BPSO, ACA, the population size of all algorithms is set to 30 and the number of iterations is set to 300. In DGGA, uniform crossover operation, directional mutation operation and elite selection strategy are used; the crossover probability p c = 0.8, and the mutation probability p m = 0.15. In BPSO, set W = 0.8, C 1 = C 2 = 1.5. In ACA, the pheromone trace is set to α = 1.0, the heuristic factor β =1.0, the volatile factor ρ = 0.7.

5.2. Experimental Results

5.2.1. Performance Comparison of Four Algorithms under Different Gain-Cost Models Using Real Datasets

We used Book and Flight datasets and generated a total of eight instances based on four gain-cost models. IGGA, DGGA, BPSO, ACA were used to obtain the best value, the worst value, the mean value, the standard deviation (S.D), and the time for solving each instance 30 times independently. Time represents the average running time required for each algorithm to solve each instance separately.
As can be seen from Table 6, IGGA achieved the best results in five of the eight instances, DGGA did that in two instances, BPSO achieved the best results on only one instance, while ACA did not. Regarding the average running time, the solution speed of DGGA and IGGA were almost equal, significantly faster than BPSO and ACA. The difference between the speed of the BPSO and ACA was small.
Since the heuristic algorithm is a random approximation algorithm, to evaluate its performance, we used the G A P metric in [40] to evaluate the average performance statistics of all algorithms. Specifically, G A P represents the relative difference between the best value and the mean value, i.e., G A P = | b e s t - m e a n | b e s t × 100 % . It can compare the average performance of all algorithms by fitting curves. If the G A P curve is closer to the abscissa axis, the average performance of the algorithm is better. From Figure 4a,b, it can be seen that among the four algorithms, the average performance of IGGA is the best, because the G A P curve is the closest to the abscissa axis.
Moreover, in order to evaluate the stability of the algorithm, we drew a histogram based on the value of S.D and evaluated the stability of all algorithms by the distribution of the columns. Figure 5a,b show that the stability of IGGA and DGGA was roughly equal in all algorithms, but significantly higher than BPSO and ACA.

5.2.2. Performance Comparison of Algorithms under Different Source Scales Using Synthetic Datasets

To investigate the scalability of the algorithm, we adapted the synthetic dataset, and evaluated the performance of IGGA through extensive experimentation. According to literature [37], we divided the dataset into two parts. The instances labeled 1–12 as small-S, and the instances labeled 13–20 as large-S. The IGGA and DGGA solutions are shown in Table 7.
The best value, the mean value, the worst value and the S.D were collected for IGGA and DGGA over 30 independent runs, which are tabulated in Table 7.
As can be seen in Table 7, for 12 small scale problem instances, IGGA had an advantage over DGGA, and there were six values superior to DGGA in the mean measure, and no difference for the remaining tests. The problem set appeared to be less challenging as both algorithms were able to reach optima in most of the cases with little difficulties.
If the quality of the IGGA solution was not surprising at the small-scale problem instance; let’s look at the performance of the large-scale problem. For instance 13–20, the IGGA has demonstrated an overwhelming advantage over DGGA. The worst solution discovered by IGGA was even better than the best solution obtained by DGGA, e.g., (14, 15, 17, 18, 20). In addition, in view of stability and consistency in finding optima, the proposed algorithm clearly surpassed DGGA with its smaller standard deviations for all problem instances. In short, IGGA has demonstrated better performance when solving source selection problems than DGGA.

5.2.3. Convergence Analysis

We have chosen two representative instances, (i.e., instance 11 and instance 16), and we also plot the solution quality of the two algorithms, as shown in Figure 6a,b. DGGA converged to local optima very quickly. However, IGGA started off slowly at the early stage of evolution due to the following two reasons. Firstly, the evolutionary process of IGGA needed to sort the ratio of gain and cost according to the early quick sorting algorithm. Secondly, our greedy strategy required traversal of all data sources, all of which required time. After undergoing crossover and mutation operations, GRS further explored other unknown areas to improve performance and generate better solutions. It can be seen that the IGGA solution was more effective than DGGA. This was the same for small-scale instance 11 and large-scale instance 16.
In fact, for large-scale instances, IGGA was not materially impacted by the large gain and cost factors, and we compared the gain-cost ratio, which simplified the problem to a new set of instances. Meanwhile, it was more in line with the demand of the solution in a practical application, so as to obtain a balance between solution quality and convergence rate.

5.2.4. Comparison with Other State-of-the-Art Models

Table 8 shows the experimental comparison of IGGA for 10 small-scale instances with previous research work. Among them, the bold represents the best optimization result, and the dash (—) indicates that the item had not been tested. In this table, we do not list the comparison results of large-scale instances, because the instance problems were randomly generated, and different experiment settings had different instance problem sets. It can be seen that IGGA was highly competitive with the most advanced methods from Table 8. IGGA can find the best value of the problem, except for instance 6, and it is especially worth mentioning that in instance 8, it found a new optimal optimization value.

6. Conclusions and Future Work

This paper studies the data source selection problem of maximizing revenue. We used data completeness, redundancy, accuracy and data coverage as evaluation indicators to establish a gain-cost model and proposed an improved greedy genetic algorithm. The selection operation and crossover operation have been improved. Finally, we conducted extensive experimental evaluations on real data set and synthetic data. The results have shown that our algorithm has efficient performance and wide applicability. The proposed method provided a new idea for data source selection.
In the future work, first, we will consider an efficient and comprehensive method to estimate data quality, propose a more complete data quality evaluation system, and select these new measures. The second step is to develop more effective source selection methods. The last is to establish a complex gain-cost model. When the data quality is multidimensional, the revenue model can be more complex.

Author Contributions

Conceived and designed the experiments, J.Y.; methodology, J.Y. and C.X.; performed the experiments/wrote the paper, J.Y.; supervision and funding acquisition, C.X.

Funding

This research was supported in part by National Nature Science Foundation of China (Grant no. 91646202).

Acknowledgments

The authors are deeply thankful to the editor and reviewers for their valuable suggestions to improve the quality of this manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Factual. Available online: https://www.factual.com/ (accessed on 5 January 2019).
  2. Infochimps. Available online: http://www.infochimps.com/ (accessed on 5 January 2019).
  3. Xignite. Available online: http://www.xignite.com/ (accessed on 27 December 2018).
  4. Microsoft Windows Azure Marketplace. Available online: https://azuremarketplace.microsoft.com/en-us/ (accessed on 8 January 2019).
  5. Dong, X.L.; Saha, B.; Srivastava, D. Less is More: Selecting Sources Wisely for Integration. Proc. VLDB Endow. 2012, 6, 37–48. [Google Scholar] [CrossRef]
  6. Rekatsinas, T.; Dong, X.L.; Srivastava, D. Characterizing and Selecting Fresh Data Sources. In Proceedings of the 2014 ACM SIGMOD International Conference on Management of Data (SIGMOD ’14), Snowbird, UT, USA, 22–27 June 2014; pp. 919–930. [Google Scholar]
  7. Lin, Y.; Wang, H.; Zhang, S.; Li, J.; Gao, H. Efficient Quality-Driven Source Selection from Massive Data Sources. J. Syst. Softw. 2016, 118, 221–233. [Google Scholar] [CrossRef]
  8. Martello, S.; Pisinger, D.; Toth, P. Dynamic Programming and Strong Bounds for the 0–1 Knapsack Problem. Manag. Sci. 1999, 45, 414–424. [Google Scholar] [CrossRef]
  9. Martello, S.; Pisinger, D.; Toth, P. New Trends in Exact Algorithms for the 0-1 Knapsack Problem. Eur. J. Oper. Res. 2000, 123, 325–332. [Google Scholar] [CrossRef]
  10. Chand, J.; Deep, K. A Modified Binary Particle Swarm Optimization for Knapsack Problems. Appl. Math. Comput. 2012, 218, 11042–11061. [Google Scholar]
  11. He, L.; Huang, Y. Research of ant colony algorithm and the application of 0-1 knapsack. In Proceedings of the 2011 6th International Conference on Computer Science & Education (ICCSE), Singapore, 3–5 August 2011; pp. 464–467. [Google Scholar]
  12. Goldberg, D.E. Genetic Algorithms in Search, Optimization and Machine Learning; Addison-Wesley: Hoboken, NJ, USA, 1989. [Google Scholar]
  13. Zhao, J.; Huang, T.; Pang, F.; Liu, Y. Genetic Algorithm Based on Greedy Strategy in the 0-1 Knapsack Problem. In Proceedings of the 2009 Third International Conference on Genetic and Evolutionary Computing, Guilin, China, 14–17 October 2009; pp. 105–107. [Google Scholar]
  14. Kallel, L.; Naudts, B.; Rogers, A. Theoretical Aspects of Evolutionary Computing; Springer-Verlag: Berlin, Germany, 2013. [Google Scholar]
  15. Drias, H.; Khennak, I.; Boukhedra, A. A hybrid genetic algorithm for large scale information retrieval. In Proceedings of the 2009 IEEE International Conference on Intelligent Computing and Intelligent Systems, Shanghai, China, 20–22 November 2009; pp. 842–846. [Google Scholar]
  16. Salomon, R. Improving the Performance of Genetic Algorithms through Derandomization. Softw. Concepts Tools 1997, 18, 175–199. [Google Scholar]
  17. Balakrishnan, R.; Kambhampati, S. Factal: Integrating Deep Web Based on Trust And relevance. In Proceedings of the 20th International Conference on World Wide Web, WWW 2011, Hyderabad, India, 28 March–1 April 2011; pp. 181–184. [Google Scholar]
  18. Calì, A.; Straccia, U. Integration of Deepweb Sources: A Distributed Information Retrieval Approach. In Proceedings of the 7th International Conference on Web Intelligence, Mining and Semantics, Amantea, Italy, 19–22 June 2017; pp. 1–4. [Google Scholar]
  19. Nuray, R.; Can, F. Automatic Ranking of Information Retrieval Systems Using Data Fusion. Inf. Process. Manag. 2006, 42, 595–614. [Google Scholar] [CrossRef]
  20. Xu, J.; Pottinger, R. Integrating Domain Heterogeneous Data Sources Using Decomposition Aggregation Queries. Inf. Syst. 2014, 39, 80–107. [Google Scholar] [CrossRef]
  21. Moreno-Schneider, J.; Martínez, P.; Martínez-Fernández, J.L. Combining heterogeneous sources in an interactive multimedia content retrieval model. Expert Syst. Appl. 2017, 69, 201–213. [Google Scholar] [CrossRef]
  22. Abel, E.; Keane, J.; Paton, N.W.; Fernandes, A.A.A.; Koehler, M.; Konstantinou, N.; Cesar, J.; Rios, C.; Azuan, N.A.; Embury, S.M. User Driven Multi-Criteria Source Selection. Inf. Sci. 2018, 431, 179–199. [Google Scholar] [CrossRef]
  23. Al Mashagba, E.; Al Mashagba, F.; Nassar, M.O. Query Optimization Using Genetic Algorithms in the Vector Space Model. arXiv, 2011; arXiv:1112.0052. [Google Scholar]
  24. Araujo, L.; Pérez-Iglesias, J. Training a Classifier for the Selection of Good Query Expansion Terms with a Genetic Algorithm. In Proceedings of the IEEE Congress on Evolutionary Computation, Barcelona, Spain, 18–23 July 2010. [Google Scholar]
  25. Sun, M.; Dou, H.; Li, Q.; Yan, Z. Quality Estimation of Deep Web Data Sources for Data Fusion. Procedia Eng. 2012, 29, 2347–2354. [Google Scholar] [CrossRef]
  26. Nguyen, H.Q.; Taniar, D.; Rahayu, J.W.; Nguyen, K. Double-Layered Schema Integration of Heterogeneous XML Sources. J. Syst. Softw. 2011, 84, 63–76. [Google Scholar] [CrossRef]
  27. Lebib, F.Z.; Mellah, H.; Drias, H. Enhancing Information Source Selection Using a Genetic Algorithm and Social Tagging. Int. J. Inf. Manag. 2017, 37, 741–749. [Google Scholar] [CrossRef]
  28. Kumar, R.; Singh, S.K.; Kumar, V. A Heuristic Approach for Search Engine Selection in Meta-Search Engine. In Proceedings of the International Conference on Computing, Communication & Automation, Noida, India, 15–16 May 2015; pp. 865–869. [Google Scholar]
  29. Abououf, M.; Mizouni, R.; Singh, S.; Otrok, H.; Ouali, A. Multi-worker multitask selection framework in mobile crowd sourcing. J. Netw. Comput. Appl. 2019. [Google Scholar] [CrossRef]
  30. Larr Naga, P.; Kuijpers, C.M.H.; Murga, R.H.; Inza, I.; Dizdarevic, S. Genetic Algorithms for the Travelling Salesman Problem: A Review of Representations and Operators. Artif. Intell. Rev. 1999, 13, 129–170. [Google Scholar] [CrossRef]
  31. Lim, T.Y.; Al-Betar, M.A.; Khader, A.T. Taming the 0/1 Knapsack Problem with Monogamous Pairs Genetic Algorithm. Expert Syst. Appl. 2016, 54, 241–250. [Google Scholar] [CrossRef]
  32. Quiroz-Castellanos, M.; Cruz-Reyes, L.; Torres-Jimenez, J.; Claudia Gómez, S.; Huacuja, H.J.F.; Alvim, A.C.F. A Grouping Genetic Algorithm with Controlled Gene Transmission for the Bin Packing Problem. Comput. Oper. Res. 2015, 55, 52–64. [Google Scholar] [CrossRef]
  33. Sarma, A.D.; Dong, X.L.; Halevy, A. Data Integration with Dependent Sources. In Proceedings of the 14th International Conference on Extending Database Technology, Uppsala, Sweden, 21–24 March 2011. [Google Scholar]
  34. Vetrò, A.; Canova, L.; Torchiano, M.; Minotas, C.O.; Iemma, R.; Morando, F. Open Data Quality Measurement Framework: Definition and Application to Open Government Data. Gov. Inf. Q. 2016, 33, 325–337. [Google Scholar] [CrossRef]
  35. Liu, W.; Liu, G. Genetic Algorithm with Directional Mutation Based on Greedy Strategy for Large-Scale 0–1 Knapsack Problems. Int. J. Adv. Comput. Technol 2012, 4, 66–74. [Google Scholar] [CrossRef]
  36. The Book and Flight datasets. Available online: http://lunadong.com/fusionDataSets.htm (accessed on 13 October 2018).
  37. Zou, D.; Gao, L.; Li, S.; Wu, J. Solving 0-1 Knapsack Problem by a Novel Global Harmony Search Algorithm. Appl. Soft Comput. J. 2011, 11, 1556–1564. [Google Scholar] [CrossRef]
  38. Pospichal, P.; Schwarz, J.; Jaros, J. Parallel Genetic Algorithm Solving 0/1 Knapsack Problem Running on the GPU. Mendel 2010, 1, 64–70. [Google Scholar]
  39. Srinivas, M.; Patnaik, L.M. Adaptive Probabilities of Crossover and Mutation in Genetic Algorithms. IEEE Trans. Syst. Man Cybern. 1994, 24, 656–667. [Google Scholar] [CrossRef]
  40. He, Y.; Wang, X. Group theory-based optimization algorithm for solving knapsack problems. Knowl.-Based Syst. 2018. [Google Scholar] [CrossRef]
  41. Zhang, X.; Huang, S.; Hu, Y.; Zhang, Y.; Mahadevan, S.; Deng, Y. Solving 0-1 Knapsack Problems Based on Amoeboid Organism Algorithm. Appl. Math. Comput 2013, 219, 9959–9970. [Google Scholar] [CrossRef]
Figure 1. Crossover operation on two selected chromosomes.
Figure 1. Crossover operation on two selected chromosomes.
Symmetry 11 00273 g001
Figure 2. Performance comparison of four gain-cost models with 35 combinations (Book dataset).
Figure 2. Performance comparison of four gain-cost models with 35 combinations (Book dataset).
Symmetry 11 00273 g002
Figure 3. Performance comparison of four gain-cost models with 35 combinations (Flight dataset).
Figure 3. Performance comparison of four gain-cost models with 35 combinations (Flight dataset).
Symmetry 11 00273 g003
Figure 4. G A P fitting curve of two data sets under four gain-cost models.
Figure 4. G A P fitting curve of two data sets under four gain-cost models.
Symmetry 11 00273 g004
Figure 5. Standard deviation (S.D) histogram of two data sets under four gain-cost models.
Figure 5. Standard deviation (S.D) histogram of two data sets under four gain-cost models.
Symmetry 11 00273 g005
Figure 6. Sample convergence plots for problem instances 11 and 16.
Figure 6. Sample convergence plots for problem instances 11 and 16.
Symmetry 11 00273 g006
Table 1. Metric definitions, description and calculation.
Table 1. Metric definitions, description and calculation.
AttributesMetricDescriptionVariablesFormula
CompletenessProportion of complete cellsIndicates the proportion of complete cells in a dataset. It means the cells that are not empty and have a meaningful value assigned. n r : Number of rows
n c : Number of columns
i c : Number of incomplete cells
n c l : Number of cells
n c l = n r n c
p c c = 1− i c n c l
RedundancyProportion of duplicate recordsRedundancy expresses the proportion of duplicate records in the data source. Since this factor is the cost-indicator, we convert it to the benefit-indicator. n r : Number of rows
r e d : Number of duplicate records
p d c = 1− r e d n r
AccuracyProportion of accurate cellsIndicate the proportion cells in a data source that has correct values according to the domain and the type of information of the data source. n c e : Number of cells with errors
n c l : Number of cells
p a c = 1− n c e n c l
Table 2. Gain-cost models.
Table 2. Gain-cost models.
CostLinearStep
Gain
LinearL-LL-S
StepS-LS-S
Table 3. Employee information provided by one data source.
Table 3. Employee information provided by one data source.
IDNameAgeSexStatusEntry_TimeSalaryCityArea Code
1Renata23Fmarried2013/550kShanghai021
3Jeremy36married2010/795kBeijing020
4ArmandFsingle2011/6Beijing010
6Leona46Mmarried120kGuangzhou021
9Renata34Fmarried2013/550kShanghai021
Table 4. Example for greedy repair strategy (GRS).
Table 4. Example for greedy repair strategy (GRS).
Source s 1 s 2 s 3 s 4 s 5 s 6
Flag111011
The ratio of gain-cost64321.51
Gain9080751004510
Cost152025503010
Table 5. The 35 different combinations of crossover and mutation probabilities.
Table 5. The 35 different combinations of crossover and mutation probabilities.
ID ( p c , p m ) ID ( p c , p m ) ID ( p c , p m ) ID ( p c , p m ) ID ( p c , p m )
1 ( 0.1 , 0.01 ) 8 ( 0.3 , 0.01 ) 15 ( 0.5 , 0.01 ) 22 ( 0.7 , 0.01 ) 29 ( 0.9 , 0.01 )
2 ( 0.1 , 0.05 ) 9 ( 0.3 , 0.05 ) 16 ( 0.5 , 0.05 ) 23 ( 0.7 , 0.05 ) 30 ( 0.9 , 0.05 )
3 ( 0.1 , 0.1 ) 10 ( 0.3 , 0.1 ) 17 ( 0.5 , 0.1 ) 24 ( 0.7 , 0.1 ) 31 ( 0.9 , 0.1 )
4 ( 0.1 , 0.3 ) 11 ( 0.3 , 0.3 ) 18 ( 0.5 , 0.3 ) 25 ( 0.7 , 0.3 ) 32 ( 0.9 , 0.3 )
5 ( 0.1 , 0.5 ) 12 ( 0.3 , 0.5 ) 19 ( 0.5 , 0.5 ) 26 ( 0.7 , 0.5 ) 33 ( 0.9 , 0.5 )
6 ( 0.1 , 0.7 ) 13 ( 0.3 , 0.7 ) 20 ( 0.5 , 0.7 ) 27 ( 0.7 , 0.7 ) 34 ( 0.9 , 0.7 )
7 ( 0.1 , 0.9 ) 14 ( 0.3 , 0.9 ) 21 ( 0.5 , 0.9 ) 28 ( 0.7 , 0.9 ) 35 ( 0.9 , 0.9 )
Table 6. Performance comparison of improved greedy genetic algorithm (IGGA), DGGA, BPSO and ACA.
Table 6. Performance comparison of improved greedy genetic algorithm (IGGA), DGGA, BPSO and ACA.
DatasetsIndexGainCostAlgorithmBestMeanWorstS.DTime
Book1LinearLinearIGGA986.32963.16921.416.842.423
DGGA953.16921.84886.6213.892.218
BPSO911.64863.71838.5014.223.135
ACA935.55896.39865.0416.493.352
2StepIGGA1092.211041.62989.589.581.883
DGGA967.08894.21847.3311.981.684
BPSO988.36918.48843.2416.692.418
ACA1013.14968.74921.3712.992.350
3StepLinearIGGA2052.451988.171923.628.732.045
DGGA2037.221942.461890.548.941.976
BPSO2055.861968.041904.8815.832.724
ACA2011.371904.321852.5618.482.831
4StepIGGA1926.051868.741831.817.171.928
DGGA1902.281828.791787.039.891.882
BPSO1908.451803.861762.2613.922.281
ACA1895.321789.921723.0417.342.292
Flight1LinearLinearIGGA436.77417.62394.176.350.116
DGGA421.21396.45376.358.060.102
BPSO413.58388.39369.6711.200.185
ACA405.18384.35374.587.470.206
2StepIGGA465.26437.72421.285.580.108
DGGA472.16431.31415.847.920.112
BPSO442.04406.46381.2610.520.168
ACA416.37390.61373.928.960.179
3StepLinearIGGA1105.541062.48970.1411.640.227
DGGA1045.34991.67966.388.950.198
BPSO992.32936.72893.1510.400.294
ACA1019.93958.61902.9413.940.315
4StepIGGA964.15908.23870.667.880.206
DGGA972.37902.52862.1710.950.192
BPSO911.63858.35810.8114.750.308
ACA902.46852.58803.9615.560.336
Table 7. The best value, the worst value, mean optima, and S.D for comparison between IGGA and DGGA.
Table 7. The best value, the worst value, mean optima, and S.D for comparison between IGGA and DGGA.
Algorithm IGGA DGGA
IDScaleBestWorstMeanS.DBestWorstMeanS.D
1102952952950295288292.321.24
22010241024102401024102410240
343535350352834.581.56
4423232302323230
515481.07481.07481.070481.07438.24478.579.38
6105050500504046.81.96
77107107107010793106.21.38
8239776977697760976797549766.320.76
9513013013001301301300
102010251025102501025102510250
113057865380576532.6357165245556054.47
124049943857490318.2549943548374235.82
13100698368546938.2316.92687967816857.3420.57
1420010,79910,67710,731.2223.4210,52210,28210,358.7336.45
1530013,36813,10413,271.9454.0412,56612,36212,472.4575.54
1650018,11117,68417,878.7492.3417,92117,45017,743.37146.08
1780037,33836,89237,172.17114.1635,14734,74134,914.81189.47
18100064,84764,31564,533.04136.3760,13258,99459,628.06285.26
19120086,39985,84786,142.4188.1786,00385,22485,647.24352.06
201500101,556101,214101,427.2664.3498,86297,93898,575.67465.23
Table 8. Comparing the experimental results with other state-of-the-art approaches.
Table 8. Comparing the experimental results with other state-of-the-art approaches.
IDInstanceIGGAMopGA [31]Harmony Algorithm [37]AOA [41]
110295295295295
220102410241024
3435353535
4423232323
515481.07481.07481.07
61050525052
77107107107107
823977697679767
95130130130130
1020102510251025

Share and Cite

MDPI and ACS Style

Yang, J.; Xing, C. Data Source Selection Based on an Improved Greedy Genetic Algorithm. Symmetry 2019, 11, 273. https://0-doi-org.brum.beds.ac.uk/10.3390/sym11020273

AMA Style

Yang J, Xing C. Data Source Selection Based on an Improved Greedy Genetic Algorithm. Symmetry. 2019; 11(2):273. https://0-doi-org.brum.beds.ac.uk/10.3390/sym11020273

Chicago/Turabian Style

Yang, Jian, and Chunxiao Xing. 2019. "Data Source Selection Based on an Improved Greedy Genetic Algorithm" Symmetry 11, no. 2: 273. https://0-doi-org.brum.beds.ac.uk/10.3390/sym11020273

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop