Next Article in Journal
A Study of Fall Detection in Assisted Living: Identifying and Improving the Optimal Machine Learning Method
Next Article in Special Issue
Network Attack Classification in IoT Using Support Vector Machines
Previous Article in Journal / Special Issue
Digital Twin-Driven Decision Making and Planning for Energy Consumption
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

OPriv: Optimizing Privacy Protection for Network Traffic

Department of Electrical and Computer Engineering, American University of Beirut, Beirut 1107 2020, Lebanon
*
Author to whom correspondence should be addressed.
J. Sens. Actuator Netw. 2021, 10(3), 38; https://0-doi-org.brum.beds.ac.uk/10.3390/jsan10030038
Submission received: 9 May 2021 / Revised: 15 June 2021 / Accepted: 21 June 2021 / Published: 24 June 2021
(This article belongs to the Special Issue Machine Learning in IoT Networking and Communications)

Abstract

:
Statistical traffic analysis has absolutely exposed the privacy of supposedly secure network traffic, proving that encryption is not effective anymore. In this work, we present an optimal countermeasure to prevent an adversary from inferring users’ online activities, using traffic analysis. First, we formulate analytically a constrained optimization problem to maximize network traffic obfuscation while minimizing overhead costs. Then, we provide OPriv, a practical and efficient algorithm to solve dynamically the non-linear programming (NLP) problem, using Cplex optimization. Our heuristic algorithm selects target applications to mutate to and the corresponding packet length, and subsequently decreases the security risks of statistical traffic analysis attacks. Furthermore, we develop an analytical model to measure the obfuscation system’s resilience to traffic analysis attacks. We suggest information theoretic metrics for quantitative privacy measurement, using entropy. The full privacy protection of OPriv is assessed through our new metrics, and then through extensive simulations on real-world data traces. We show that our algorithm achieves strong privacy protection in terms of traffic flow information without impacting the network performance. We are able to reduce the accuracy of a classifier from 91.1% to 1.42% with only 0.17% padding overhead.

1. Introduction

Nowadays, statistical traffic analysis is becoming an attractive tool for developing algorithms to evaluate and manage internet network traffic. By definition, traffic analysis is a network engineering technique that consists of examining statistical features of flow packets (e.g., packet sizes, inter-arrival times, and packet directions) and building classifiers, using machine-learning algorithms to infer traffic information. Network administrators take advantage of traffic analysis techniques for the purposes of intrusion detection to detect abnormal traffic [1,2,3], and to differentiate malicious traffic flows [4]. In addition, a proper deployment of traffic analysis provides valuable insights for resource management, traffic control, diagnostic checking, and provisioning. Engineers can use this information to build robust networks and avoid possible delays. To this end, traffic analysis is used to support internet-based services, including banking, health, military, government, electrical systems, and transportation.
On the other hand, traffic analysis can be misused to launch attacks to infer knowledge on network traffic by means of exploiting side-channel information leakage. Even with the growing use of encryption to protect traffic content, traffic analysis is becoming a common attack, threatening privacy, anonymity, and confidentiality. These types of attacks can cause major privacy breaches on military systems, banking, healthcare, etc.
In fact, an adversary can misuse traffic analysis in order to conclude a user’s online activity, which is typically private and includes sensitive information. These privacy breaches, also referred to as ‘side-channel information leaks’ range from predicting users’ locations [5], distinguishing the downloaded web pages [6]; identifying language in an encrypted VoIP conversation [7]; obtaining records of an encrypted VoIP chat [8]; or identifying critical information about the underlying type of applications [9,10,11].
In brief, traffic analysis nowadays is considered a major threat of confidentiality of network applications. Additionally, tools for protecting information traffic privacy still fall short of what is required in terms of strong privacy protection and low performance overhead. In this work, we aim to answer the following question: can we optimally protect users’ privacy without deteriorating internet traffic behavior?
This work is an extension and improved version of our previous work [12] to fully address the problem of optimizing network obfuscation. We aim for a traffic obfuscation system that strikes a good balance between efficiency in obstructing application tracing and maintaining minimal performance impact. Our goal is to find the sufficient padding or fragmentation that guarantees good secrecy while optimizing the overhead. In addition, we present a structured and comprehensive overview into entropy-based obfuscation systems in order to ensure security and efficiency at the same time. The main contributions of this work are as follows:
  • Verification of the benefits achieved by an optimal solution for the mathematical problem of network traffic obfuscation problem;
  • Suggestion of information theoretic metrics for quantitative privacy measurement, using entropy;
  • Proposal of OPriv, a practical solution based on the IBM ILOG CPLEX optimization package (ILOG);
  • Verification of the effectiveness of OPriv via two criteria, using both analytical and experimental models;
  • Conducting a comparative study between OPriv and state-of-the-art obfuscation models.
The rest of this paper is organized as follows. We summarize the related work in Section 2. Section 3 presents the threat model, the design of the obfuscation system, and the experiments to evaluate the efficiency of the system. In Section 4, we present new metrics based on entropy to enhance obfuscation systems. In Section 5, we suggest a measure to evaluate our proposed obfuscation model and compare it with other techniques. Section 6 discusses the implication issues. Finally, we conclude the paper in Section 7.

2. Literature

The study of traffic analysis for network data flow is not new. Given the importance of traffic classification for purposes of security and quality of service (QoS) management, different machine learning solutions have been used effectively [2,13,14,15,16,17].
Conversely, there is currently a new research direction focusing on traffic obfuscation to thwart classification in order to protect users’ security. Different systems have been developed to ensure confidentiality, performance, and efficiency of networks. In the following, we group the classification-obfuscation methods into four types: anonymization, mutation, morphing, and tunneling. We classify some of the reviewed work in Table 1.
Anonymization consists of hiding key information that provides important information for traffic classification, such as IP addresses, port numbers, and MAC addresses. In this context, multi-path routing [29,30] and NATing [31] can be used for anonymizing the communicated traffic. TOR is a well-known system for anonymous communication based on the second generation of the onion routing model [32]. However, in the recent past, it was verified that even such systems are vulnerable to detection by machine-learning techniques [33,34,35] which have proved efficient in classifying TOR traffic. Other proposed anonymization techniques consist of inserting additional dummy packets and transmitting them to conceal the real traffic. For instance, ref. [18] proposed the use of heavy traffic to hinder the adversary’s ability to tamper with the links. Additionally, the proposed defenses in ref. [36,37,38] deliberately drop some packets, known as defensive dropping, and in ref. [39], inject artificial delay intentionally.
Traffic mutation relies on changing the flows’ statistical characteristics to confuse a classifier and to make it difficult to identify the original traffic [40]. It consists of properly modifying the packet sizes and/or the packet interarrival times (IAT), using padding, fragmentation, and buffering; consequently, the statistics of the overall conveyed traffic become considerably dissimilar from the original one. In this context, padding and fragmentation are used to hide packet size information. On the other hand, buffering aims to hide the interarrival time information. These methods have a great impact in reducing the accuracy of statistically-based traffic classifiers. Maximum transmission unit (MTU) padding, for example, is a well-known mutation technique that consists of padding all the flow packets to the maximum payload size MTU [19]. Random padding is another technique that consists of randomly padding the packets [41].
However, mutation and anonymization methods can result in a drastic overhead in terms of the amount of data being sent [42]. This leads to performance degradation in terms of delay and bandwidth consumption [43]. On the other hand, some works proved that it is still possible to classify the traffic, even when implementing these methods [44].
Morphing techniques aim at confusing the classifier into classifying the target application traffic as another type. Wright et al. proposed a morphing technique that consists of transforming one class of traffic to look like another class by applying convex optimization [20,21]. Nevertheless, this method applies to devices with strong computational capabilities. In addition, it increases latency, due to the generation of random numbers for each input packet.
Tunneling hides packet-related features by the use of encryption and the creation of virtual networks. The use of VPN, for example, ensures hiding the connection metadata and, consequently, the users’ privacy. However, the work in [45] categorizes VPN traffic based on the application name and traffic type. Ref. [22] suggests using virtualized MAC addresses to obscure the adversary’s analysis. Their model consists of creating multiple virtual MAC interfaces over a single wireless card. Then, the packets over these interfaces are dynamically scheduled, and the packet features are reshaped over each virtual interface. However, this solution is more appropriate for using a Wi-Fi connection on a computer.
We proposed in our previous works, so far, a handful of obfuscation algorithms resulting in significant outcomes [46,47,48,49]. Our goal was to address the privacy protection of network communication against side information based inferential attacks by exploring a combination of concepts, specifically anonymization, mutation, and morphing.
In ref. [12], we presented a mathematical model for network obfuscation. Then, we formulated analytically the problem of selecting both the target app from a data set and the packet length from the designated target app to mutate to.
Despite the efficient results of our earlier obfuscation algorithms, we aim in this work to analyze mathematically our previous results to find a systematic explanation for them. This can be realized using concepts of information theory. Information entropy is a mathematical concept that measures how much information there is in an event [50]. It was widely implemented in security systems to detect traffic anomalies [51,52], or for systems of DDoS defense [53]. To the best of our knowledge, the concept of measure function that helps in explaining and predicting theoretically the obfuscation results for a specific defense system has not been tackled yet in the literature. In this work, we aim to provide a better understanding of the implementation of entropy-based methods in obfuscation systems against traffic analysis attacks.

3. Traffic Privacy Protection

3.1. Adversary Model

We consider a passive adversary who uses machine-learning methods to trace end-user application identification through network traffic classification. In wireless LANs, eavesdropping is easy, due to the shared medium; hence, encrypted traffic samples that users send over wireless links are practically exposed to sniffers monitoring the traffic. Even worse, the attacker can have direct access to the encrypted data of a user from the server. This adversary could be, for example, a government that gained access to an ISP. Hence, the adversary has flow features information, such as timing, size, direction, and count of packets in a specific encrypted network flow.
We illustrate in the reference scenario model in Figure 1, an adversary who monitors the internet connection of a mobile user. The adversary uses sniffer software (e.g., Wireshark) and does not have any knowledge about the software or encryption schemes that the user implements. The classification system collects traffic traces of the user and identifies which mobile network application they are running. We use this traffic classification attack scenario to test our obfuscation system model in Section 3.5.

3.2. Classification Attack

We consider six popular network applications, including chatting, online gaming, video calling and streaming, VoIP, and social media browsing. We list the applications and their differences in Table 2, and we label them from 1 to 6. We build four machine learning algorithms as explained in our previous work [46]. The accuracy of the different classifiers and their corresponding parameters are presented in Table 3.

3.3. Obfuscation Optimization Problem

Information leakage of the flow content is provided typically from traffic features, such as packet lengths, inter-packet times, traffic direction, etc.
To achieve maximum protection, we formulated mathematically in [12] the optimization problem with the goal of maximizing obfuscation, while minimizing the overhead and being subject to traffic constraints. We divide the packet length traces in each app traffic set into bins of size w. Then, we study the probability of each bin of packet lengths across all instances. We modify each packet in the flows, created by the source app to defend, such that their distribution becomes similar to packet sizes of a different target app.
Consider an application i defined by [ m i , L i , P i ] where the following is true:
  • m i is the number of possible packet length ranges of size w.
  • L i = [ L i 1 , L i 2 , …, L i m i ] is the vector of smallest lengths in each bin of size w. The average of L i is denoted by Avg( L i ).
  • P i = [ P i 1 , P i 2 ,.., P i m i ] is the vector of length probabilities for bins of size w.
An incoming packet from the source application can be defined by its size l and the probability p of the bin to which l belongs. We define a binary variable for each application that indicates whether the app is selected as target app or not; x i denotes the variable for the ith app. Given a set of N applications, the vector X (of length N) represents [ x 1 , x 2 , …, x N ].
In addition, we delineate a binary variable for each bin of an application i that indicates whether the packet bin is selected to be mutated to or not. It is worth noting that the smallest size in a bin is used for mutation. y i j denotes the variable for the jth bin of application i. Given a set of m i ranges, the vector y i (of length m i ) designates [ y i 1 , y i 2 , …, y i m i ].
In order to minimize the accuracy of the classifier, we need to find the probability of a packet length in a bin of a target app such that it is similar to p. At the same time, to minimize the overhead, the selected packet length must be similar to l. In the case that the difference between p and the designated probability is too small (less than 0.1 in our case), the difference between the selected length and l needs to be checked that it is less than the average difference. This guarantees that the overhead does not reach a large value.
The obfuscation optimization problem for mutating an incoming packet from a source app is as follows: Find the vector X (of length N), and the vector y i (of length m i ), and minimize:
M i n i = 1 N x i j = 1 m i y i j ( P i j p 2 × L i j l 2 )
Subject to the following:
1.
P i j [ 0 , 1 ] ; i = 1 , . . . , N
2.
L i j l < A v g ( L i ) l f o r P i j p < 0.1
3.
i = 1 N x i = 1 ; x i = 0 o r 1 ; i = 1 , . . . , N
4.
i [ 1 , N ] , j = 1 m i y i j = 1 ; y i j = 0 o r 1 ; j = 1 , . . . , m i

3.4. OPriv

In this part, we present OPriv, a practical solution to the optimization problem. At a high-level, given a source application, our goal is to mutate a source application traffic such that it is confused as some different target application. The user chooses first the source application that they would like to defend, as well as a set of target applications that they would like to make the source application resemble. Of course, we need to make sure that the implemented mutation results in a minimum added number of overhead bytes. OPriv dictates how each packet size from the source application should be mutated such that the resulting distribution resembles the target with a minimum of overhead.
The model has constraints, including a quadratic term and therefore, it is a mixed-integer quadratically constrained program (MIQCP) problem. As a practical result, we deploy and solve the problem deciding on mixed-integer quadratic problem MIQP linearization, using the software IBM ILOG CPLEX Optimization Studio Version: 12.9.0.0. CPLEX is chosen for its robustness in solving large-scale, real-world problems. The coding in CPLEX specifies the problem parameters, optimization variables, objective function, and constraints discussed previously. The result for one iteration is as follows: vector X to identify the target app for mutation, and the corresponding Y to specify the packet bin to mutate to in the target app.
The proposed method, implemented using IBM’s CPLEX optimization package (ILOG), requires 1–4 min to solve with a dual-core Intel i7 processor.
To solve several instances of the same problem, we define our model in Java and call Cplex Java API on Eclipse to solve it. We record in Table 4 the time needed to execute each instance of optimization for the different applications.

3.5. Obfuscation Experiments and Results

For each incoming packet from the source app, we implement the solution of the optimization problem to select the optimal target app and packet size to mutate to. First, we select the unique values of lengths and their corresponding probabilities of a certain app in our data set as input to our code in IBM ILOG CPLEX. The output results are the optimal packet sizes to mutate to. We generate tables mapping each length source packet to its corresponding optimal target size offline, and so the entire process becomes more practical, inducing minimal latency, especially for dynamic applications. Those mapping tables are the results of our optimization problem and define how we should mutate packets. Skype, Facebook and Game have around 1400 unique values of sizes. WhatsApp and Viber have around 200 unique values. As for YouTube, the number is 1000. For the sake of simplicity, we show only a sample of seven mapping values in Table 5, Table 6, Table 7, Table 8, Table 9 and Table 10. According to those tables, the difference between the original length and the optimal length to mutate to is minimal, which achieves our goal in minimizing the padding overhead. For each incoming packet, OPriv refers to the appropriate mapping table, matches its packet length and probability using our offline tables and infers the corresponding optimum packet size that we need to mutate to. Finally, we apply our masking by means of padding and fragmenting. If the incoming packet length is less than the corresponding optimum length, we proceed by padding it with zeroes such that the resultant packet length become similar. Otherwise, we chop the packet into two smaller fragments where one of them has the same size as the corresponding optimum length.
To evaluate OPriv, we use our four model classifiers described in Section 3.2, and we predict the classes of the mutated traffic. We present, in Table 11, comprehensive simulation results of the mutation using OPriv of each application in our data set. For example, mutating Game using OPriv can decrease the SVM classifier’s accuracy from 76.7% to 0.85%, Bagged Trees classifier’s accuracy from 90% to 1.45%, KNN classifier’s accuracy from 83.9% to 1.83%, and Random Forest classifier’s accuracy from 91.1% to 1.42% with only 0.17% average overhead. As expected, using optimization, OPriv outperforms all of the other state-of-the-art defense solutions against traffic analysis. It realizes a good balance between efficiency and overhead, especially for Game, WhatsApp, Viber, and YouTube. We explain in what follows the inefficiency of Skype and Facebook mutation.

4. Entropy Concepts to Enhance Obfuscation Efficiency

Shannon entropy and Kullback–Leibler divergence (also known as information divergence or relative entropy) are perhaps the two most important quantities in information theory and its applications. Most of these measures have never found any application in machine learning and network obfuscation systems. In an attempt to explain the efficiency of our obfuscation systems, we present in what follows the metrics of information to test their performance, namely, degree of traffic masking effectiveness, Kullback–Leibler divergence, and traffic divergence.

4.1. Degree of Traffic Masking Effectiveness

Given a random variable X, which takes values of the finite set of M values x 1 , x 2 ,…, x M and p i := P(X = x i ) is the probability of occurrence of x i ; hence, the Shannon entropy is as follows:
H = i = 1 M p i l o g ( p i )
The conditional entropy of a variable Y given X of M values is defined by the following:
H ( Y / X ) = i = 1 M p ( X = i ) H ( Y / X = i )
Consider a traffic network and its set of p features X = X 1 , X 2 , …, X p , the degree of traffic masking effectiveness using a certain obfuscation algorithm that mutates any of the p features, can be studied in terms of conditional entropy.
Most traffic analysis attacks use information derived from the sequences of packet lengths, and/or from interarrival times as input. This motivates our definition of degree of traffic masking effectiveness to measure how much the adversary can learn about a certain network application, given the observed packet lengths or gap times leaked by the traffic.
Let Φ be the set of packets that refer to a specific application, and let Ψ be the side information on Φ gained by the adversary by observing the encrypted traffic. The degree of masking effectiveness is measured by the conditional entropy H ( Φ / Ψ ) to denote how much information an adversary can learn about the identification of a certain application, given the observed traffic features. Accordingly, the smaller that H ( Φ / Ψ ) is, the more likely that Φ is a function of Ψ . This means that mutating or modifying any pattern of Ψ sincerely affects the identification of Φ . Moreover, a high H ( Φ / Ψ ) indicates that Φ is less dependent on Ψ . Hence, no matter how much the obfuscation algorithm mutates Ψ , that does not significantly affect the fingerprinting of Φ .
As a proof of concept, we study the degree of the masking effectiveness of H ( Φ / Ψ ) for the apps in our data set. We evaluate H ( Φ / Ψ ) for variables of side information being packet lengths, and packet interarrival time IAT. We present the results in Table 12.
According to Table 12, the values of H ( Φ / Ψ ) with Ψ = packet lengths, are relatively low, indicating a high dependency of our apps on their packet lengths. In this case, any security algorithm that mutates the lengths effectively will obfuscate the adversary knowledge of the exact application, using traffic analysis. The impact of mutation of lengths of Game, WhatsApp, Viber, and YouTube are foreseen and tangible, as H ( Φ / Ψ ) for those apps are relatively minor in comparison to those of Skype and Facebook. This can explain the results in Table 11, where OPriv appears to be less efficient in the cases of Skype and Facebook, in comparison to the other apps.
The values of H ( Φ / Ψ ) with Ψ = packet IAT are relatively high. This means that in our case of traffic traces, the knowledge of the exact type of application does not mainly depend on IATs. Any obfuscation algorithm that modifies the IAT would be less effective in obscuring the application identity.
In our previous work [47], we proposed AdaptiveMutate, a privacy thwarting technique with 3 variations where we mutate the packet lengths, and/ or interarrival times of the source app to defend such that the lengths or IAT of the output packets appear as though they are coming from the target app probability mass function. The simulation results show a significant efficiency of AdaptiveMutate when applied using lengths mutation. However, IAT mutation is not efficient enough in reducing the classifier’s accuracy. This is proved by the values of H ( Φ / Ψ ) in Table 12.

4.2. Kullback–Leibler Divergence

In our previous works [49], we presented an obfuscation algorithm that consists of mutating the packet lengths of a source app to those lengths from a target app, having similar bin probability. We used a trial-and-error process where we mutated comprehensively, using our algorithm, each application to all the other apps in order to decide on the best target app to mutate to. We present instead in this part the Kullback–Leibler divergence, a mathematical defined metric to decide on which target app is best to mutate to.
Given two complete discrete probability distributions P = ( p 1 , p 2 , …, p n ), and Q = ( q 1 , q 2 , …, q n ), where p i , q i ∈ [0,1] and i = 1 n p i = i = 1 n q i = 1. The Kullback–Leibler divergence is a measure of divergence between P and Q. According to [54], the Kullback–Leibler D ( P Q ) is obtained using the following formula:
D ( P Q ) = i = 1 n ( p i l o g ( p i q i ) )
The Kullback–Leibler divergence can be used as a decision function to decide on which network application in a data set to mutate to. In our context of network obfuscation assessment, we denote by P the probability distribution of a feature in the source app, and by Q that of the destination app to mutate to. The smaller D ( P Q ) is, the closer the distributions P and Q are, and vice versa. A better obfuscation efficiency would be realized in cases where Q is chosen in such a way that D ( P Q ) is large. The more that the source and destination app feature probability distributions are different, the more there is changeable features of the source app, and consequently, it would confuse it as the destination app by the classifier. The changeable features make the accurate identification of network application more difficult, whereas, in the case that P and Q are similar, the mutation of P would not change much its distribution, making the confusion less significant.
Table 13 presents KL divergence for lengths probabilities of the different apps and explains the results of our previous works [49] where mutation to target apps with higher values of KL lead to better obfuscation results.
According to Table 14, KL divergence for IAT probabilities estimation of the different apps are close to zero and therefore, are very similar. Hence, there is no changeable features in the case of IAT mutation, making the confusion of a classifier less significant. This is proved in the results of [49] when we try to mutate IATs.

4.3. Traffic Divergence

Traffic divergence T D ( P Q ) between the original and mutated traffic can be expressed in terms of Kullback–Leibler divergence. P denotes in this case the probability distribution of a certain side information before mutation, and Q that after mutation. Hence, T D ( P Q ) measures to what extent a certain obfuscation algorithm can modify the probability distribution of a side information of an app traffic. T D reflects the efficiency of the obfuscation algorithm, and on the amount of overhead resulting from it. A high T D indicates a high change of the probability distribution of the specified side information feature. If this feature brings a small degree of traffic masking (detailed in Section 4.1), that results in an efficient protection against traffic analysis. However, a high traffic divergence can be reflected in a higher overhead. T D can be used to compare between obfuscation algorithms. It gives an insight of how much the algorithm causes modification and overhead. The finest obfuscation algorithm is the one that mutates side information with the smallest degree of traffic masking and results in least T D at the same time. Table 15 presents the traffic divergence between the original traffic of our app data set, and the mutated traffic, using OPriv.

5. Defense Evaluation

We propose in this section a measure to evaluate obfuscation techniques against traffic analysis attacks. This measure describes how good a certain obfuscation model is and compares between different defense systems. We name it Obfuscation Effectiveness (OE) and we define it using two metrics: the attacker confidence, and the data overhead.
The attacker confidence is defined as the estimated ratio of correct app inferences to attempted app inferences by an attacker with no previous knowledge if an obfuscation technique is already deployed. Lesser attacker confidence indicates that the technique is more efficient at defending user privacy. With no protection, attack confidence can reach 1.
Data overhead reflects the padding overhead induced by adding additional dummy bytes by a certain obfuscation technique. It is defined as the ratio of network data sent with and without a certain obfuscation technique. For instance, a bandwidth overhead of 1.5 indicates that applying the technique results in 1.5 times as much bytes traffic sent on unprotected network traffic. This ratio is generally greater than 1. Of course, a smaller data overhead is desirable because additional traffic leads to more network congestion.
A main issue for the robustness of an obfuscation technique is to check how much effort it requires, and how much overhead it induces at the same time. Eventually, we define the obfuscation effectiveness as the product of attacker confidence and the data overhead. Some obfuscation models have a fixed product but others are adjustable to users’ preferences for privacy versus data expenses.
In this context, obfuscation effectiveness is a measure that reflects the model performance as a function of the amount of overhead and secrecy realized. The smaller the obfuscation effectiveness, the more the defense model has good performance in terms of secrecy realized and the resulting data overhead.
We present in Table 16, the obfuscation effectiveness of the Random Forest classifier for some obfuscation algorithms suggested in our previous work, in comparison with OPriv. As expected, in using optimization, OPriv has the smallest value of obfuscation effectiveness among all the other techniques and therefore has the best results.

6. Discussion and Future Work

The work presented in this study is realistic and precise to implement in real-world practice. It is differential from other past studies by the way that OPriv is coded in an optimization software so that it is immediate and practical. In addition, the mapping of values between original and mutated is realized as length to length and not bin to bin. This implies better precision of our algorithm than those of previous works. The results show the effectiveness of our obfuscation system in balancing performance and security. It ensures privacy protection and reduces bandwidth overhead at the same time. Therefore, OPriv is suitable for all types of network devices and does not need many processing requirements.
Another advantage of our method is that the tables mapping the mutation needed to modify the incoming length into the optimal length can be realized offline. Thus, OPriv can run in a dynamic online manner with minimal latency.
The mapping of the packet lengths from a source app to the corresponding optimal lengths in the target app, Table 5, Table 6, Table 7, Table 8, Table 9 and Table 10, and the obfuscation results in Table 11 are very compatible with the same mapping and results in our previous work [12]. This evidences the solidarity of the analytical model in [12] without the need for an optimization software. However, it makes it easier to use an available commercial software.
The information metrics presented in this research proved, for the first time, the obfuscation results of our work and of other previous works. This gives significant understanding of the obfuscation results for any thwarting technique. In addition, these metrics can serve as criteria for selecting tunable parameters to achieve the best results of the obfuscation model.

7. Conclusions

This paper addresses the privacy protection of network communication against side information based inferential attacks. We study optimal methods to secure network traffic under the scrutiny of an adversary who uses statistical traffic analysis. We present an effective approach, OPriv, to thwart traffic classification optimally: we characterize leaking information, model the attack mathematically, study a heuristic to solve the optimization problem, and evaluate it, using both analytical and experimental models.
Our practical method OPriv is based on solving the nonlinear optimization problem with the IBM ILOG optimization package. The simulation results based on real data traffic confirm that OPriv has reasonable performance in reducing the success rate of statistical traffic analysis attacks. The simulation verifies that this proposed solution performs well on a real data set by reducing the classification accuracy from 91.1% to 1.42% with only 0.17% padding overhead. We also provide in this work novel metrics based on entropy to evaluate, for any obfuscation system, the right choice of features to mutate, as well as the right choice of target applications to mutate to. The mathematical modeling, information theory evaluation, and experimental results presented offer both theoretic and real-world proofs to protect network users’ privacy.

Author Contributions

Formal analysis, L.C.; Methodology, L.C.; Supervision, A.C. and A.K.; Writing—original draft, L.C.; Writing—review & editing, A.C.and A.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the AUB University Research Board, and TELUS Corp., Canada.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data underlying this article will be shared on reasonable request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Callegari, C.; Coluccia, A.; D’Alconzo, A.; Ellens, W.; Giordano, S.; Mandjes, M.; Pagano, M.; Pepe, T.; Ricciato, F.; Zuraniewski, P. A methodological overview on anomaly detection. Data Traffic Monit. Anal. 2013, 7754, 148–183. [Google Scholar]
  2. Aloqaily, M.; Otoum, S.; Al Ridhawi, I.; Jararweh, Y. An intrusion detection system for connected vehicles in smart cities. Hoc Netw. 2019, 90, 101842. [Google Scholar] [CrossRef]
  3. Buczak, A.L.; Guven, E. A survey of data mining and machine learning methods for cyber security intrusion detection. IEEE Commun. Surv. Tutor. 2015, 18, 1153–1176. [Google Scholar] [CrossRef]
  4. Pasqualetti, F.; Dörfler, F.; Bullo, F. Attack detection and identification in cyber-physical systems. IEEE Trans. Autom. Control 2013, 58, 2715–2729. [Google Scholar] [CrossRef] [Green Version]
  5. Das, A.K.; Pathak, P.H.; Chuah, C.N.; Mohapatra, P. Privacy-aware contextual localization using network traffic analysis. Comput. Netw. 2017, 118, 24–36. [Google Scholar] [CrossRef] [Green Version]
  6. Wang, T. High precision open-world website fingerprinting. In Proceedings of the 2020 IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 18–20 May 2020; pp. 152–167. [Google Scholar]
  7. Wang, C.; Kennedy, S.; Li, H.; Hudson, K.; Atluri, G.; Wei, X.; Sun, W.; Wang, B. Fingerprinting encrypted voice traffic on smart speakers with deep learning. In Proceedings of the 13th ACM Conference on Security and Privacy in Wireless and Mobile Networks, Linz, Austria, 8–10 July 2020; pp. 254–265. [Google Scholar]
  8. White, A.M.; Matthews, A.R.; Snow, K.Z.; Monrose, F. Phonotactic reconstruction of encrypted voip conversations: Hookt on fon-iks. In Proceedings of the 2011 IEEE Symposium on Security and Privacy, Oakland, CA, USA, 22–25 May 2011; pp. 3–18. [Google Scholar]
  9. Wang, Q.; Yahyavi, A.; Kemme, B.; He, W. I know what you did on your smartphone: Inferring app usage over encrypted data traffic. In Proceedings of the 2015 IEEE conference on communications and network security (CNS), Florence, Italy, 28–30 September 2015; pp. 433–441. [Google Scholar]
  10. Taylor, V.F.; Spolaor, R.; Conti, M.; Martinovic, I. Appscanner: Automatic fingerprinting of smartphone apps from encrypted network traffic. In Proceedings of the 2016 IEEE European Symposium on Security and Privacy (EuroS&P), Saarbruecken, Germany, 21–24 March 2016; pp. 439–454. [Google Scholar]
  11. Yamansavascilar, B.; Guvensan, M.A.; Yavuz, A.G.; Karsligil, M.E. Application identification via network traffic classification. In Proceedings of the 2017 International Conference on Computing, Networking and Communications (ICNC), Santa Clara, CA, USA, 26–29 January 2017; pp. 843–848. [Google Scholar]
  12. Chaddad, L.; Chehab, A.; Elhajj, I.H.; Kayssi, A. Optimal Packet Camouflage against Traffic Analysis. ACM Trans. Priv. Secur. 2021, in press. [Google Scholar]
  13. Pacheco, F.; Exposito, E.; Gineste, M.; Baudoin, C.; Aguilar, J. Towards the deployment of machine learning solutions in network traffic classification: A systematic survey. IEEE Commun. Surv. Tutor. 2018, 21, 1988–2014. [Google Scholar] [CrossRef] [Green Version]
  14. D’Alconzo, A.; Drago, I.; Morichetta, A.; Mellia, M.; Casas, P. A survey on big data for network traffic monitoring and analysis. IEEE Trans. Netw. Serv. Manag. 2019, 16, 800–813. [Google Scholar] [CrossRef] [Green Version]
  15. Salman, O.; Elhajj, I.H.; Kayssi, A.; Chehab, A. A review on machine learning–based approaches for internet traffic classification. Ann. Telecommun. 2020, 75, 673–710. [Google Scholar] [CrossRef]
  16. Velan, P.; Čermák, M.; Čeleda, P.; Drašar, M. A survey of methods for encrypted traffic classification and analysis. Int. J. Netw. Manag. 2015, 25, 355–374. [Google Scholar] [CrossRef]
  17. Dhote, Y.; Agrawal, S.; Deen, A.J. A survey on feature selection techniques for internet traffic classification. In Proceedings of the 2015 International Conference on Computational Intelligence and Communication Networks (CICN), Jabalpur, India, 12–14 December 2015; pp. 1375–1380. [Google Scholar]
  18. Fu, X.; Graham, B.; Bettati, R.; Zhao, W. On countermeasures to traffic analysis attacks. In Proceedings of the IEEE Systems, Man and Cybernetics SocietyInformation Assurance Workshop, West Point, NY, USA, 18–20 June 2003; pp. 188–195. [Google Scholar]
  19. Dyer, K.P.; Coull, S.E.; Ristenpart, T.; Shrimpton, T. Peek-a-boo, i still see you: Why efficient traffic analysis countermeasures fail. In Proceedings of the 2012 IEEE Symposium on Security and Privacy, Francisco, CA USA, 21–23 May 2012; pp. 332–346. [Google Scholar]
  20. Qu, B.; Zhang, Z.; Guo, L.; Zhu, X.; Guo, L.; Meng, D. An empirical study of morphing on network traffic classification. In Proceedings of the 7th International Conference on Communications and Networking in China, Kunming, China, 8–10 August 2012; pp. 227–232. [Google Scholar]
  21. Wright, C.V.; Coull, S.E.; Monrose, F. Traffic Morphing: An Efficient Defense Against Statistical Traffic Analysis. In NDSS; Citeseer; Internet Society: San Diego, CA, USA, 8–11 February 2009; Volume 9. [Google Scholar]
  22. Zhang, F.; He, W.; Chen, Y.; Li, Z.; Wang, X.; Chen, S.; Liu, X. Thwarting Wi-Fi side-channel analysis through traffic demultiplexing. IEEE Trans. Wirel. Commun. 2013, 13, 86–98. [Google Scholar] [CrossRef] [Green Version]
  23. Juarez, M.; Imani, M.; Perry, M.; Diaz, C.; Wright, M. Toward an efficient website fingerprinting defense. In European Symposium on Research in Computer Security; Springer: Berlin/Heidelberg, Germany, 2016; pp. 27–46. [Google Scholar]
  24. Imani, M.; Rahman, M.S.; Wright, M. Adversarial traces for website fingerprinting defense. In Proceedings of the 2018 ACM SIGSAC Conference on Computer and Communications Security, Toronto, ON, Canada, 15–19 October 2018; pp. 2225–2227. [Google Scholar]
  25. Lu, D.; Bhat, S.; Kwon, A.; Devadas, S. Dynaflow: An efficient website fingerprinting defense based on dynamically-adjusting flows. In Proceedings of the 2018 Workshop on Privacy in the Electronic Society, Toronto, ON, Canada, 15–19 October 2018; pp. 109–113. [Google Scholar]
  26. Sirinam, P.; Imani, M.; Juarez, M.; Wright, M. Deep fingerprinting: Undermining website fingerprinting defenses with deep learning. In Proceedings of the 2018 ACM SIGSAC Conference on Computer and Communications Security, Toronto, ON, Canada, 15–19 October 2018; pp. 1928–1943. [Google Scholar]
  27. Liberatore, M.; Levine, B.N. Inferring the source of encrypted HTTP connections. In Proceedings of the 13th ACM Conference on Computer and Communications Security, Alexandria, VA, USA, 30 October–3 November 2006; pp. 255–263. [Google Scholar]
  28. Panchenko, A.; Niessen, L.; Zinnen, A.; Engel, T. Website fingerprinting in onion routing based anonymization networks. In Proceedings of the 10th Annual ACM Workshop on Privacy in the Electronic Society, Chicago, IL, USA, 17 October 2011; pp. 103–114. [Google Scholar]
  29. Liu, H.; Wang, Z.; Miao, F. Concurrent multipath traffic impersonating for enhancing communication privacy. Int. J. Commun. Syst. 2014, 27, 2985–2996. [Google Scholar] [CrossRef]
  30. Hu, Y.; Li, X.; Liu, J.; Ding, H.; Gong, Y.; Fang, Y. Mitigating Traffic Analysis Attack in Smartphones with Edge Network Assistance. In Proceedings of the 2018 IEEE International Conference on Communications (ICC), Kansas City, MO, USA, 20–24 May 2018; pp. 1–6. [Google Scholar]
  31. Gokcen, Y.; Foroushani, V.A.; Heywood, A.N.Z. Can we identify NAT behavior by analyzing Traffic Flows? In Proceedings of the 2014 IEEE Security and Privacy Workshops, San Jose, CA, USA, 17–18 May 2014; pp. 132–139. [Google Scholar]
  32. Syverson, P.; Dingledine, R.; Mathewson, N. Tor: The secondgeneration onion router. In Security Symposium; USENIX Association: San Diego, CA, USA, 9–13 August 2004; pp. 303–320. [Google Scholar]
  33. Yuan, Z.; Wang, C. An improved network traffic classification algorithm based on Hadoop decision tree. In Proceedings of the 2016 IEEE International Conference of Online Analysis and Computing Science (ICOACS), Chongqing, China, 28–29 May 2016; pp. 53–56. [Google Scholar]
  34. Hodo, E.; Bellekens, X.; Iorkyase, E.; Hamilton, A.; Tachtatzis, C.; Atkinson, R. Machine learning approach for detection of nontor traffic. In Proceedings of the 12th International Conference on Availability, Reliability and Security, Reggio Calabria, Italy, 29 August–1 September 2017; pp. 1–6. [Google Scholar]
  35. Montieri, A.; Ciuonzo, D.; Bovenzi, G.; Persico, V.; Pescapé, A. A dive into the dark web: Hierarchical traffic classification of anonymity tools. IEEE Trans. Netw. Sci. Eng. 2019, 7, 1043–1054. [Google Scholar] [CrossRef] [Green Version]
  36. Shmatikov, V.; Wang, M.H. Timing analysis in low-latency mix networks: Attacks and defenses. In European Symposium on Research in Computer Security; Springer: Berlin/Heidelberg, Germany, 2006; pp. 18–33. [Google Scholar]
  37. Fu, X.; Graham, B.; Bettati, R.; Zhao, W.; Xuan, D. Analytical and empirical analysis of countermeasures to traffic analysis attacks. In Proceedings of the 2003 International Conference on Parallel Processing, Kaohsiung, Taiwan, 6–9 October 2003; pp. 483–492. [Google Scholar]
  38. Wang, W.; Motani, M.; Srinivasan, V. Dependent link padding algorithms for low latency anonymity systems. In Proceedings of the 15th ACM conference on Computer and Communications Security, Alexandria, VA, USA, 27–31 October 2008; pp. 323–332. [Google Scholar]
  39. Levine, B.N.; Reiter, M.K.; Wang, C.; Wright, M. Timing attacks in low-latency mix systems. In International Conference on Financial Cryptography; Springer: Berlin/Heidelberg, Germany, 2004; pp. 251–265. [Google Scholar]
  40. Cheng, T.H.; Lin, Y.D.; Lai, Y.C.; Lin, P.C. Evasion techniques: Sneaking through your intrusion detection/prevention systems. IEEE Commun. Surv. Tutor. 2011, 14, 1011–1020. [Google Scholar] [CrossRef] [Green Version]
  41. Diyanat, A.; Khonsari, A.; Shariatpanahi, S.P. A dummy-based approach for preserving source rate privacy. IEEE Trans. Inf. Forensics Secur. 2016, 11, 1321–1332. [Google Scholar] [CrossRef]
  42. Iacovazzi, A.; Baiocchi, A. Investigating the trade-off between overhead and delay for full packet traffic privacy. In Proceedings of the 2013 IEEE International Conference on Communications Workshops (ICC), Budapest, Hungary, 9–13 June 2013; pp. 1345–1350. [Google Scholar]
  43. Iacovazzi, A.; Baiocchi, A. From ideality to practicability in statistical packet features masking. In Proceedings of the 2012 8th International Wireless Communications and Mobile Computing Conference (IWCMC), Limassol, Cyprus, 27–31 August 2012; pp. 456–462. [Google Scholar]
  44. Fu, X.; Graham, B.; Bettati, R.; Zhao, W. On effectiveness of link padding for statistical traffic analysis attacks. In Proceedings of the 23rd International Conference on Distributed Computing Systems, Providence, RI, USA, 19–22 May 2003; pp. 340–347. [Google Scholar]
  45. Wang, W.; Zhu, M.; Wang, J.; Zeng, X.; Yang, Z. End-to-end encrypted traffic classification with one-dimensional convolution neural networks. In Proceedings of the 2017 IEEE International Conference on Intelligence and Security Informatics (ISI), Beijing, China, 22–24 July 2017; pp. 43–48. [Google Scholar]
  46. Chaddad, L.; Chehab, A.; Elhajj, I.H.; Kayssi, A. App traffic mutation: Toward defending against mobile statistical traffic analysis. In Proceedings of the IEEE INFOCOM 2018-IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS), Honolulu, HI, USA, 15–19 April 2018; pp. 27–32. [Google Scholar]
  47. Chaddad, L.; Chehab, A.; Elhajj, I.H.; Kayssi, A. AdaptiveMutate: A technique for privacy preservation. Digit. Commun. Netw. 2019, 5, 245–255. [Google Scholar] [CrossRef]
  48. Chaddad, L.; Chehab, A.; Elhajj, I.H.; Kayssi, A. Mobile traffic anonymization through probabilistic distribution. In Proceedings of the 2019 22nd Conference on Innovation in Clouds, Internet and Networks and Workshops (ICIN), Paris, France, 19–21 February 2019; pp. 242–248. [Google Scholar]
  49. Chaddad, L.; Chehab, A.; Elhajj, I.H.; Kayssi, A. Network Obfuscation for Net Worth Security. In Proceedings of the 2020 Seventh International Conference on Software Defined Systems (SDS), Paris, France, 30 June–3 July 2020; pp. 83–88. [Google Scholar]
  50. Shannon, C.E. A mathematical theory of communication. ACM Sigmobile Mob. Comput. Commun. Rev. 2001, 5, 3–55. [Google Scholar] [CrossRef]
  51. Han, W.; Xue, J.; Yan, H. Detecting anomalous traffic in the controlled network based on cross entropy and support vector machine. IET Inf. Secur. 2019, 13, 109–116. [Google Scholar] [CrossRef] [Green Version]
  52. Yang, C. Anomaly network traffic detection algorithm based on information entropy measurement under the cloud computing environment. Clust. Comput. 2019, 22, 8309–8317. [Google Scholar] [CrossRef]
  53. Kalkan, K.; Altay, L.; Gür, G.; Alagöz, F. JESS: Joint entropy-based DDoS defense scheme in SDN. IEEE J. Sel. Areas Commun. 2018, 36, 2358–2372. [Google Scholar] [CrossRef]
  54. Kullback, S.; Leibler, R.A. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
Figure 1. Mobile traffic classification attack.
Figure 1. Mobile traffic classification attack.
Jsan 10 00038 g001
Table 1. Traffic obfuscation techniques.
Table 1. Traffic obfuscation techniques.
WorkYearAttacker ModelAcc. BeforeDefense MethodAcc. AfterDefense Approach
[18]2003Bayes100%Dummy packets insertion50%Anonymization
[19]2012Liberatore and Levine87%MTU padding41%Mutation
[19]2012Herrmann et al.98%Random padding40%Mutation
[19]2012Herrmann et al.98%Random MTU padding11%Mutation
[19]2012Herrmann et al.98%Linear padding73%Mutation
[19]2012Herrmann et al.98%Exponential padding61%Mutation
[20]2012Naive Bayes65%Elephants and mice padding28%Mutation
[21]2009Naive Bayes98%Convex Optimization63%Morphing
[22]2013SVM and NN83%Traffic demultiplexing44%Other
[23]2016K-NN91%Adaptive padding20%Mutation
[24]2018CNN model98%Random padding60%Mutation
[25]2018Optimal attacker94%DynaFlow44%Morphing and Mutation
[26]2018Deep learning90%BuFLO12%Mutation
[27]2006Naive Bayes98%MTU padding7%Mutation
[28]2011SVM80%Camouflage4%Anonymization
Table 2. Experiment data set  [49].
Table 2. Experiment data set  [49].
#AppType
1SkypeVideo call
2FacebookInteractive browsing
38 ball poolGame
4WhatsAppMessaging
5ViberVOIP
6YouTubeVideo streaming
Table 3. Data set accuracy  [49].
Table 3. Data set accuracy  [49].
ModelParametersAccuracy
SVMQuadratic kernel76.7%
Bagged treesMinimum leaf size = 290.0%
KNNNumber of neighbors = 5;83.9%
Distance weight = squared inverse
Random forestMinimum leaf size = 291.1%
Table 4. Average execution time for each instance of optimization for the different applications.
Table 4. Average execution time for each instance of optimization for the different applications.
ApplicationAverage Execution Time (s)
Skype0.16
Facebook0.36
Game0.2
WhatsApp0.14
Viber0.36
YouTube0.19
Table 5. Sample of length mapping of Skype App, using OPriv.
Table 5. Sample of length mapping of Skype App, using OPriv.
Original LengthOriginal Length ProblemOptimum Length to Mutate to
850.008654
1030.0072154
1200.008154
1340.00623154
1450.004064104
7600.00031804
14400.0004311354
Table 6. Sample of length mapping of Facebook app using OPriv.
Table 6. Sample of length mapping of Facebook app using OPriv.
Original LengthOriginal Length ProblemOptimum Length to Mutate to
660.3353954
740.00487104
1220.00015154
1870.00030205
3840.000335320
8980.00014854
15140.272811504
Table 7. Sample of length mapping of Game App, using OPriv.
Table 7. Sample of length mapping of Game App, using OPriv.
Original LengthOriginal Length ProblemOptimum Length to Mutate to
740.01562104
780.01908154
1020.00757355
5440.00286704
3630.00113554
7320.00105804
12100.001861307
Table 8. Sample of length mapping of WhatsApp app, using OPriv.
Table 8. Sample of length mapping of WhatsApp app, using OPriv.
Original LengthOriginal Length ProblemOptimum Length to Mutate to
660.43874104
740.01267254
1920.00694154
2920.00815304
3080.00724355
9140.006631057
15140.267051354
Table 9. Sample of length mapping of Viber app, using OPriv.
Table 9. Sample of length mapping of Viber app, using OPriv.
Original LengthOriginal Length ProblemOptimum Length to Mutate to
500.0628154
630.04657104
650.04173304
1500.02118205
1550.02219254
1600.02028154
5920.00119654
Table 10. Sample of length mapping of YouTube app, using OPriv.
Table 10. Sample of length mapping of YouTube app, using OPriv.
Original LengthOriginal Length ProblemOptimum Length to Mutate to
660.3338162
780.01202104
5000.00562592
5140.00668240
5480.00616300
14840.01667804
15140.438931354
Table 11. Results of apps mutation, using OPriv.
Table 11. Results of apps mutation, using OPriv.
AlgorithmSVMBTKNNRFOverhead
Original76.7%90%83.9%91.1%-
Skype40.90%51.21%43.68%40.92%3.45 %
Facebook42.2%47.77%47.54%41.12%4.41%
Game0.85%1.45%1.83%1.42%0.17%
WhatsApp11.25%12.93%10.8 %10.83%0.93%
Viber13.2%12.58%11.74%11.78%0.84%
YouTube8.97%2.58%8.49%9.1%1.58%
Table 12. Degree of traffic masking effectiveness.
Table 12. Degree of traffic masking effectiveness.
Ψ LengthsIAT
Φ
Skype0.37760.9336
Facebook0.43740.7545
Game0.11680.7530
WhatsApp0.04510.3511
Viber0.18620.6020
YouTube0.26350.4362
Table 13. Kullback–Leibler divergence of packet lengths probabilities.
Table 13. Kullback–Leibler divergence of packet lengths probabilities.
QSkypeFacebookGameWhatsAppViberYouTube
P
Skype01.77752.59472.51882.22432.6258
Facebook1.962004.26583.14264.28313.2296
Game2.32624.172803.08782.33562.1204
WhatsApp2.93863.12403.105703.86564.2221
Viber2.66132.73773.68443.147102.3251
YouTube2.60193.16972.12574.23742.87940
Table 14. Kullback–Leibler divergence of IAT probabilities.
Table 14. Kullback–Leibler divergence of IAT probabilities.
QSkypeFacebookGameWhatsAppViberYouTube
P
Skype00.00060.00020.02260.06550.0017
Facebook0.000900.00180.01940.06170.0006
Game0.00020.000900.02310.06650.0019
WhatsApp0.07270.05120.079700.05000.0324
Viber0.28630.21540.35290.07300.1933
YouTube0.00330.00110.00390.01580.05920
Table 15. Length traffic divergence T D between original and mutated traffic, using OPriv.
Table 15. Length traffic divergence T D between original and mutated traffic, using OPriv.
AppLength TD
Skype4.9948
Facebook6.5112
Game3.2843
WhatsApp4.2030
Viber3.8317
YouTube4.8685
Table 16. Obfuscation effectiveness (OE).
Table 16. Obfuscation effectiveness (OE).
Obfuscation ModelVersionOE
AdaptiveMutate [47]Packet length mutation0.369
AdaptiveMutate [47]IAT mutation0.071
AdaptiveMutate [47]Lengths and IAT mutation0.084
Anonymization through prob. distr. [48]First alg. (to normal distr. of Skype)0.087
Anonymization through prob. distr. [48]First alg. (to Poisson distr. of Viber)0.137
Anonymization through prob. distr. [48]Second algorithm0.16
Mutation To similar prob. distr. [49]-0.02
Optimal Analytical Solution [12]-0.019
OPriv-0.015
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chaddad, L.; Chehab, A.; Kayssi, A. OPriv: Optimizing Privacy Protection for Network Traffic. J. Sens. Actuator Netw. 2021, 10, 38. https://0-doi-org.brum.beds.ac.uk/10.3390/jsan10030038

AMA Style

Chaddad L, Chehab A, Kayssi A. OPriv: Optimizing Privacy Protection for Network Traffic. Journal of Sensor and Actuator Networks. 2021; 10(3):38. https://0-doi-org.brum.beds.ac.uk/10.3390/jsan10030038

Chicago/Turabian Style

Chaddad, Louma, Ali Chehab, and Ayman Kayssi. 2021. "OPriv: Optimizing Privacy Protection for Network Traffic" Journal of Sensor and Actuator Networks 10, no. 3: 38. https://0-doi-org.brum.beds.ac.uk/10.3390/jsan10030038

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop