Next Article in Journal
Measuring and Energizing Sensor System for Digital Signal Monitoring of an Academic–Experimental CubeSat for Wireless Telemetry Purposes
Previous Article in Journal
Structural Optimization Design of Dual Robot Gripper Unloading Device Based on Intelligent Optimization Algorithms and Generative Design
Previous Article in Special Issue
Energy Aware Load Balancing Framework for Smart Grid Using Cloud and Fog Computing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unsupervised Mixture Models on the Edge for Smart Energy Consumption Segmentation with Feature Saliency

Concordia’s Institute for Information Systems Engineering (CIISE), Concordia University, Montreal, QC H3G 1M8, Canada
*
Author to whom correspondence should be addressed.
Submission received: 30 August 2023 / Revised: 29 September 2023 / Accepted: 5 October 2023 / Published: 7 October 2023
(This article belongs to the Special Issue Machine Learning and Data Analytics for Edge Cloud Computing)

Abstract

:
Smart meter datasets have recently transitioned from monthly intervals to one-second granularity, yielding invaluable insights for diverse metering functions. Clustering analysis, a fundamental data mining technique, is extensively applied to discern unique energy consumption patterns. However, the advent of high-resolution smart meter data brings forth formidable challenges, including non-Gaussian data distributions, unknown cluster counts, and varying feature importance within high-dimensional spaces. This article introduces an innovative learning framework integrating the expectation-maximization algorithm with the minimum message length criterion. This unified approach enables concurrent feature and model selection, finely tuned for the proposed bounded asymmetric generalized Gaussian mixture model with feature saliency. Our experiments aim to replicate an efficient smart meter data analysis scenario by incorporating three distinct feature extraction methods. We rigorously validate the clustering efficacy of our proposed algorithm against several state-of-the-art approaches, employing diverse performance metrics across synthetic and real smart meter datasets. The clusters that we identify effectively highlight variations in residential energy consumption, furnishing utility companies with actionable insights for targeted demand reduction efforts. Moreover, we demonstrate our method’s robustness and real-world applicability by harnessing Concordia’s High-Performance Computing infrastructure. This facilitates efficient energy pattern characterization, particularly within smart meter environments involving edge cloud computing. Finally, we emphasize that our proposed mixture model outperforms three other models in this paper’s comparative study. We achieve superior performance compared to the non-bounded variant of the proposed mixture model by an average percentage improvement of 7.828%.

1. Introduction

The predictive power of machine learning holds the key to deciphering intricate patterns and driving efficient solutions for a sustainable future, particularly in the realm of smart meter data modelling and utility program improvement. This predictive capability has already played a crucial role in ensuring global food supplies, a once seemingly insurmountable challenge. Scientists have been instrumental in harnessing this power for the benefit of society. As we embark on this new era, machine learning is poised to further our understanding of the world, optimize resource utilization, and reduce our environmental impact, ultimately promoting prosperity and sustainability. In this pursuit, our focus is on the intricacies of smart meter data modelling and its application in enhancing utility programs, such as energy efficiency and demand response. The implementation of Advanced Metering Infrastructure (AMI) across Europe stands as a notable catalyst behind the surpassing of energy efficiency targets outlined in the EU’s 20-20-20 energy policy. Building on the triumphs in Europe, smart meter deployments have transcended borders, becoming a global phenomenon in nations striving to modernize their electricity grids. Consequently, these groundbreaking advancements in energy metering technologies have birthed a trove of high-quality, consistently sampled electrical power consumption datasets. This surge in data dimensions underscores the compelling necessity for meticulous feature selection within the domain of machine learning models. This ensures the prioritization of the most enlightening attributes while simultaneously mitigating noise and curtailing computational expenditures. Within the machine learning context, “features” denote the distinct measurable properties or intrinsic characteristics of data that serve as the essential input for predictive models. In the scope of this paper, when we allude to “features”, we specifically refer to the statistical metrics derived from time-series data or the readings gleaned from smart meters pertaining to a particular energy consumer. Our work delves into the challenges and potential of this domain, introducing methodologies that not only improve the predictive accuracy but also enhance transparency and interpretability. Our goal is to ensure that every stakeholder, from scientists to policymakers, can fully utilize the potential of these advancements for a brighter and more sustainable future.
The challenge of smart meter data modelling using clustering techniques is pivotal in advancing utility programs geared towards achieving energy sustainability and fostering a better future. In this context, the integrated IoT architecture for smart metering proposed by the research in [1] provides valuable insights into the technological foundations of modern smart metering systems. Effectively harnessing high-frequency smart meter data to understand consumer energy consumption behaviour presents a significant opportunity. Research papers, exemplified by [2], have delved into the segmentation of household energy consumption using hourly data, enabling the identification of intricate consumption patterns. Likewise, the work in [3] revolves around the analysis and clustering of residential customers’ energy behavioural demand using smart meter data, facilitating the recognition of distinct consumption behaviours [3,4,5,6,7]. These modelling solutions offer substantial benefits by providing utility programs with tailored insights. They empower utilities to develop strategies for energy efficiency and demand response that are intricately aligned with consumer behaviour. Ultimately, this not only enhances energy sustainability but also contributes to the creation of a more environmentally responsible and prosperous future.
Moreover, the richer and more granular data may lead to more complex and diverse consumption patterns, necessitating the use of flexible distributions in statistical models to capture the nuances in class data distributions effectively [8,9,10]. DR is an incentive program that allows utility companies to save money on unnecessary investments and lower emissions of greenhouse gases (GHG) [8,9,10]. DR induces households to reduce their energy consumption levels at high wholesale market prices or when system reliability is jeopardized. EE programs aim to reduce the power demand of households while maintaining their consumption habits [8,11,12,13]. Traditional machine learning exploratory analysis tools, such as unsupervised learning techniques, transform smart meter information into valuable information participating in customer clustering [8]. Clustering is a statistical data analysis technique that can uncover or infer intrinsic properties and cluster the data into several components according to the observations’ similarities [8]. As a soft clustering approach, the Gaussian mixture’s reliability and minimal impact on computational capabilities have made it a good candidate for modelling smart meter data [8,14,15,16]. The Gaussian distribution does not fit data well within a mixture model if the data have an asymmetric distribution, as demonstrated in Figure 1. The estimation of data-bounded support regions using Gaussian mixture models has been a notable avenue of research, with advancements in vector quantization techniques [17,18,19,20,21]. The deployment of AMI has introduced high dimensionality in modern energy consumption datasets [4]. Patterns are easily distinguished within observations represented with features of high entropy. Feature selection has several advantages: it is well established to improve the performance of model-based classification [22], and it helps to develop interpretable models that are reduced in complexity within applications across several disciplines [23]. The search for the optimal number of clusters and the optimal set of features is an interrelated optimization problem [23]. However, searching for the optimal set of features is challenging in an unsupervised setting because there is no clear criterion for the optimization process, since the number of clusters is unknown [23]. Historically, to find the optimal number of features, an exhaustive search is done through the space of all feature subsets [24,25,26]. Additionally, non-exhaustive search techniques do not guarantee finding the optimal feature subset. Therefore, an efficient solution was proposed within an unsupervised setting [23]; the optimal feature subset search is converted into an estimation problem parallel to the learning of mixture models, where a vector of feature weights is estimated using the expectation-maximization (EM) algorithm [23].
In our experimental analysis, our proposed method outperforms the asymmetric generalized Gaussian mixture model-based feature selection (FSAGGMM), the bounded asymmetric generalized Gaussian mixture model (BAGGMM), and the asymmetric generalized Gaussian mixture model (AGGMM) according to several performance evaluation metrics. Additionally, our proposed mixture model has been implemented using Concordia University’s High-Performance Computing (HPC) Facility: Speed [27].
The current energy consumer segmentation approach distinguishes itself from previous works by effectively modelling different representations of smart meter data, taking into account the class data bounds, inferring the true number of consumer clusters, and finding the optimal set of features in a single optimization process. The rest of the paper is organized as follows: in Section 2, we inform the reader about all the prior works within the context of this paper. in Section 3, we describe the proposed feature selection model based on the bounded asymmetric generalized Gaussian mixture model (FSBAGGMM). Section 4 explains how the mixture model’s parameters are estimated and how the MML’s objective function is derived for our specific case. Section 5 exhibits the experimental results in the context of household energy consumption segmentation by comparing the performance of our proposed algorithm against several state-of-the-art clustering algorithms. Finally, we discuss and conclude our research in Section 6 and Section 7, respectively.

2. Prior Works

Numerous applications leverage energy consumption data, benefiting from the increased feasibility and reliability facilitated by smart meters. Non-intrusive load monitoring (NILM) has enhanced heating, ventilation, and air conditioning (HVAC) fault detection through smart meter readings, eliminating the need for additional sensors [28]. Smart meter data serve as valuable input for load forecasting and energy efficiency recommendations [29]. Customer-oriented solutions, such as user-friendly web portals for bill understanding, have also been proposed [30]. Additionally, energy consumption data inform predictive models and offer consumption insights, further contributing to energy efficiency [29,31,32]. Previous research has addressed key aspects of smart meter data analytics. The research in [33] focused on smart-meter-driven segmentation, while the research in [34] introduced layer-wise relevance propagation for smart grid stability prediction. The research in [35] optimized deep models for improved smart grid stability prediction. Additionally, the research in [36] explored customer segmentation based on smart meter data analytics. These studies form the foundation for our research, covering various aspects of smart meter data analysis and its applications.
Clustering has proven helpful to find energy consumption patterns in low- and high-voltage customers [37,38]. Additionally, demand management programs have successfully utilized clustering in order to select suitable candidate energy consumers [39,40,41]. Thus, several approaches have been employed for the segmentation of energy users, such as Euclidean distance-based clustering [31,38] and multi-resolution clustering in the spectral domain [42]. Similarly, several clustering methods, such as hierarchical clustering, K-means, fuzzy K-means, and self-organizing maps (SOM), have been used to cluster consumers with similar energy consumption patterns in [37]. SOM was tested for its capability to classify consumption profiles in [43]. Clustering has also proven useful to enhance energy consumption prediction using a two-layer feed-forward artificial neural network [10]. The Gaussian mixture model, optimized by the EM algorithm, was utilized in [32,44] as a non-distance-based consumer segmentation tool. Other finite mixture models have also been used within the context of the same application [45].
In order to model smart meter data in different representations, several limitations imposed by the Gaussian mixture model must be overcome. Several distributions have been used as a base distribution of mixture models to overcome the shape rigidity of the Gaussian distribution, such as the Student’s-t distribution [46,47,48] and the generalized Gaussian distribution (GGD) [49,50,51]. Compared to the Gaussian distribution, the Student’s-t distribution has an additional parameter ( ν ) called the degree of freedom that allows the distribution to generalize to different probability distributions. The Student’s-t distribution is identical to the Cauchy distribution when ( ν = 1 ) and approaches the Gaussian distribution as ( ν ) approaches infinity. As for the GGD, the additional parameter per component ( λ ) is called the shape parameter; it controls the tails of the distribution, making it far more flexible to different types of data and less vulnerable to outliers [52,53,54]. In more recent studies, the asymmetric generalized Gaussian distribution (AGGD) was used as a base distribution for mixture models [55,56]. The AGGD can generalize to a large class of distributions, such as the impulsive, the Laplacian, the Gaussian, and the uniform distributions, in addition to the ability to fit asymmetric data [57]. Additionally, and in order for mixture components to fit better to real-life data, the bounded support concept was adopted in several finite mixture models [17,20,58].
Several feature extraction methods have been utilized to process high-dimensional data in electrical load observations and convert them into a new set of reduced feature spaces. In [59], a scalable algorithm for data processing has been proposed for a dataset collected from 10,000 Australian homes over a year. Dimensionality reduction is accomplished by employing a sparse representation technique in [60]. An encoding system has given representations for energy consumers with a pre-processed dictionary in [2]. The discovery of prominent energy consumption time windows is crucial for feature extraction and, therefore, in modelling the typical consumer’s behaviour. Through a thorough analysis of several smart meter trials, researchers have been able to identify four time periods where the most extensive distribution of peak demand occurs within smart meter datasets [32]. The energy consumption data within the specified time periods were used to calculate seven weakly correlated features. Projection methods such as principal component analysis (PCA) were also used to concisely represent a consumer’s load curve [37].
In the context of the energy consumption segmentation application, a feature selection approach based on genetic algorithms has been utilized effectively in [31] to reduce the high dimensionality of smart meter data and improve the clustering performance of k-means. In general, several exhaustive search methods are conducted to perform feature selection, such as sequential forward search, backward search, floating search, beam search, bidirectional search, and genetic search [24,25,26,61]. However, more recently, several studies have approached the problem of finding the optimal set of features as an optimization problem within the context of mixture-based clustering in several real-life applications [56,62], thus achieving feature selection with minimal computation expenses.
Various methods have been employed to determine the optimal number of energy consumer clusters. Diverse clustering evaluation metrics and scenarios have been utilized, with the best scenario dictating the optimal number of consumption profiles [63,64]. Additionally, an entropy-based evaluation index was applied to time series data for cluster optimization [31]. Probabilistic model selection methods, such as the Bayesian Information Criterion (BIC) and the Akaike Information Criterion (AIC), were used in different studies to select the optimal cluster count [32,65]. It is worth noting that the AIC tends to favour more complex models, particularly with smaller training datasets, while the BIC leans toward simpler models. Another superior approach is the Minimum Message Length (MML) criterion, known for its excellence over BIC and AIC [66,67,68]. MML, combined with the feature-weighting mixture model [23], simultaneously performs model and feature selection, avoiding exhaustive searches. This paper builds on prior research that has evolved mixture models to become increasingly flexible and assumption-light, aiming to better capture real-world data complexities. Our proposed model leverages this accumulated knowledge to introduce a more flexible approach.

3. The Unsupervised BAGGMM-Based Feature Selection Model

Mixture models are a powerful approach to model incomplete data. The observations in this paper are represented as a set of vectors X = { X 1 , X 2 , X 3 , , X N } , X i R D , i { 1 , 2 , 3 , , N } . We aim to model data in X using a mixture model with M components where M 1 . It is possible to state that the D-dimensional random variable X i = ( X i 1 , X i 2 , , X i D ) is sampled from a M component mixture model if its probability density function can be written as follows:
p ( X i | Θ ) = k = 1 M p ( X i | θ k ) p k
where Θ represents the set of parameters of all the M-component mixture models. The term p k represents the mixing proportion of the component k; by definition, p k is positive and k = 1 M p k = 1 . The likelihood function gives the joint distribution for all the observations:
p ( X | Θ ) = i = 1 N k = 1 M p ( X i | θ k ) p k
In order to define the complete data likelihood, an M-dimensional vector of unobserved variables is defined, and it is denoted by Z i . For each observation i, the unobserved binary vector is assigned with 0 s, except at the k’th position, where the cluster is responsible primarily. The complete data likelihood is defined as follows:
p ( X , Z | Θ ) = i = 1 N k = 1 M p ( X i | θ k ) p k Z i k
where Z = { Z 1 , , Z N } . The features in Equation (2) are considered to be of equal importance. However, in the context of a real application, the estimation of the feature weights is an effective approach to better model data [37,38]. The integration of the feature selection approach within the mixture model involves considering that the irrelevant features are modelled with a background Gaussian distribution as in [23]. In this paper, feature weights are estimated for all the mixture components. Therefore, the background Gaussian distribution has a single set of parameters β = { η , δ } , where η represents the vector of means for all the data dimensions and δ represents the standard deviation vector. Thus, we are proposing to rewrite Equation (2) to adopt feature relevancy as follows:
p ( X i | Θ , β , φ ) = k = 1 M p j d = 1 D p ( X i d | θ k d ) φ d p ( X i d | β d ) 1 φ d
where β = { ( η 1 , δ 1 ) , , ( η D , δ D ) } . The unobserved binary vector φ = ( φ 1 , , φ D ) indicates the relevancy of each feature. By assuming that the elements within vector φ are mutually exclusive and independent of the component label Z, we have
p ( X i , φ ) = p ( X i | φ ) p ( φ ) = k = 1 M p k d = 1 D ω d p ( X i d | θ k d ) φ d × ( 1 ω d ) p ( X i d | β d ) 1 φ d
After the marginalization over φ , the obtained mixture model is formalized as follows:
p ( X i | Θ M ) = k = 1 M p k d = 1 D ω d p ( X i d | θ k d ) + ( 1 ω d ) p ( X i d | β d )
where Θ M = [ Θ , ω , β ] is the complete set of parameters that define the proposed mixture model. The vector ω = ( ω 1 , , ω D ) quantifies the feature importance with a set of weights where ω d = p ( φ d = 1 ) . Thus, Equation (6) represents the probability density function that is assumed to generate the data. The foreground distribution or the mixture base distribution p ( X i d | θ k d ) models the relevant attributes of each latent class in the data. Several distributions have been proposed for feature selection in the context of mixture models, such as the asymmetric Gaussian distribution (AGD) [62] and the asymmetric generalized Gaussian distribution (AGGD) [56]. However, these distributions are unbounded with a support region that extends across the set of real numbers. Real-life datasets are mostly digitized and have bounded support [18]. Therefore, we propose the bounded asymmetric generalized Gaussian distribution (BAGGD) to model the relevant features of each component in the mixture. The BAGGD distribution generalizes several different distribution classes, such as the impulsive, the Laplacian, the Gaussian, and the uniform distributions, to fit different shapes of observed bounded support, asymmetric, and non-Gaussian data. In order to define the bounded distribution proposed in this paper, the bounded support region τ k d in R for each component is first defined for the following indicator function:
H ( X i d | k ) = 1 X i d τ k d 0 O t h e r w i s e
The bounded asymmetric generalized Gaussian probability density function for each D-dimensional data point is defined as follows:
p ( X i | θ k ) = d = 1 D Ψ ( X i d | θ k d ) H ( X i d | k ) k Ψ ( X i d | θ k d ) d X
The unbounded distribution p ( X i d | θ k d ) is the asymmetric generalized Gaussian distribution (AGGD). The symmetric and asymmetric generalized Gaussian distributions are defined in Equations (9) and (10), respectively.
g ( X i d | μ k d , σ k d , λ k d ) = λ k d Γ ( 3 / λ k d ) Γ ( 1 / λ k d ) 1 / 2 2 σ k d Γ ( 1 / λ k d ) e x p A ( λ k d ) | X i d μ k d σ k d | λ k d
Ψ ( X i d | θ k d ) = g 1 ( X i d | θ k d ) x < μ k d g 2 ( X i d | θ k d ) x μ k d = λ k d Γ ( 3 / λ k d ) Γ ( 1 / λ k d ) 1 / 2 ( σ l k d + σ r k d ) Γ ( 1 / λ k d ) × e x p A ( λ k d ) μ k d X i d σ r k d λ k d X i d < μ k d e x p A ( λ k d ) X i d μ k d σ l k d λ k d X i d μ k d
where A ( λ k d ) = Γ ( 3 / λ k d ) Γ ( 1 / λ k d ) λ k d / 2 ; θ k d = [ μ k d , σ l k d , σ r k d , λ k d ] represents the set of parameters that defines the AGGD for each mixture component. μ k d , σ l k d , σ r k d , and λ k d denote the mean, the left standard deviation, the right standard deviation, and the shape parameter of the AGGD, respectively. The shape parameter controls the distribution’s tails. The larger its value, the flatter the distribution at the mean; the smaller it is, the more peaked the distribution at the mean. The right and left variance combination allows the probability density function to be asymmetric or non-asymmetric. Thus, the proposed mixture model would consider the different shapes, asymmetry, and bounded support region of the smart meter data. Bounded distribution generalizes to all its special cases, including the bounded variants [18]. Thus, our proposed FSBAGGMM generalizes to a wide range of mixture models, including the bounded variants, as shown in Table 1. Additionally, we will demonstrate in Section 5 how the proposed FSBAGGMM can generalize feature selection models based on the asymmetric generalized Gaussian mixture, in addition to several specific mixture models in terms of modelling smart meter data.

4. Model Parameter Estimation and Selection

In this section, we will explain how the feature weights and the mixture model parameters are estimated for the modelling of the training data, in addition to the model selection criterion. We propose an approach to reveal the valid number of intrinsic clusters within a dataset using MML and estimate the proposed model’s parameters using EM.

4.1. Parameter Estimation Using the EM Algorithm

The mixture model’s parameters are optimized in parallel with the features’ weights in each iteration using the EM algorithm. The iterations of the EM algorithm produce a sequence of models with a non-decreasing log-likelihood. The parameters are optimized to achieve the maximum log-likelihood, and the log-likelihood function is expressed as follows:
L ( X , Θ M , Z , φ ) = i , k p ( Z i = k | X i ) log p k + i , k d φ d = 0 1 p ( Z i = k , φ | X i ) × ( φ d ( log ( p ( X i d | θ k d ) + log w d ) + ( 1 φ d ) ( log p ( X i d | β d ) + log ( 1 ω d ) ) )
The EM algorithm has made the optimization process for mixture models feasible through an iterative process using Equation (11) instead of Equation (2). The conditional expected values γ ( Z j h ) and ω d ^ are given by Equations (12) and (13).
p ( Z i = k | X i , Θ M ) = γ ( Z i k ) = p k d = 1 D ζ i , k , d j = 1 K p j d = 1 D ζ i , j , d
ω ^ d = i = 1 N j = 1 M ω d p ( X i d | θ j d ) ζ i , j , d γ ( Z i j ) N
where ζ i , k , d = ω d p ( X i d | θ k d ) + ( 1 ω d ) p ( X i d | β d ) . The EM algorithm consists of a loop over two steps: the E-step and the M-step. They are performed repetitively until convergence. In the E-step, Equation (12) is evaluated using either the initial parameters or the parameters estimated in the M-step. In the M-step, the parameters of the next model in the sequence are estimated. Each estimated model in the sequence represents a better approximation of the distribution of the smart meter data. Due to the complicated nature of the BAGGD function, the gradient of the log-likelihood function (Equation (11)) with respect to each one of the parameters was non-linear, and a closed-form solution was not obtained; therefore, for these parameters, we used the Newton–Raphson method to approximate the update values, as demonstrated in the equations below. The partial derivatives obtained with respect to each of the parameters can be found in Appendix A. Thus, the M-step is implemented using the following equations:
p k = p ( Z k = 1 ) = i = 1 N p ( k | X i , Θ M ) N
μ k ^ d = μ k d 2 L ( X , Θ M , Z , φ ) μ k d 2 1 L ( X , Θ M , Z , φ ) μ k d
σ l ^ k d = σ l k d 2 L ( X , Θ M , Z , φ ) σ l k d 2 1 L ( X , Θ M , Z , φ ) σ l k d
σ r ^ k d = σ r k d 2 L ( X , Θ M , Z , φ ) σ r k d 2 1 L ( X , Θ M , Z , φ ) σ r k d
λ k ^ d = λ k d 2 L ( X , Θ M , Z , φ ) λ k d 2 1 L ( X , Θ M , Z , φ ) λ k d
η ^ d = i = 1 N ( 1 ω d ) p ( X i d | β d ) ζ i , k , d γ ( Z i k ) x i d i = 1 N j = 1 M ( 1 ω d ) p ( X i d | β d ) ζ i , j , d γ ( Z i j )
δ ^ d 2 = i = 1 N ( 1 ω d ) p ( X i d | β d ) ζ i , k , d γ ( Z i k ) ( x i d η d ) 2 i = 1 N j = 1 M ( 1 ω d ) p ( X i d | β d ) ζ i , j , d γ ( Z i j )

4.2. Model Selection

Model selection involves selecting the best set of parameters that model the smart meter data. Among several candidate models, the model with the maximum log-likelihood may achieve the best fit to the data; however, it is not guaranteed to perform well on unseen data. In other words, model evaluation based on the log-likelihood exclusively could be misleading. In this section, we develop a model selection criterion to infer the true number of consumption profiles within a dataset in an unsupervised manner. The Minimum Message Length criterion [72,73] is an information-theory-based model selection method; it selects the best model among a list of candidate statistical models based on its capability of compressing a message containing the data. According to the MML criterion, the best model minimizes a message that consists of two parts: the first part encodes the model using prior knowledge about the model exclusively, and the second part encodes the data using the model. Given a list of candidate models, the following function is minimized to obtain the true number of intrinsic clusters within the data:
MessLens log p ( Θ M ) + c 2 ( 1 + log ρ c ) + 1 2 log | I ( Θ M ) | log p ( X | Θ M )
In Equation (21), the prior distribution is represented by p ( Θ M ) , the determinant of the Fisher information matrix is represented by | I ( Θ M ) | , and the model’s likelihood is represented by p ( X | Θ M ) . The constant c is the total number of parameters; in this case, it is calculated as c = M + D + 4 D M + 2 D , c 1 . The term ρ c R c represents the optimal quantization lattice constant [74]; the value of the constant is approximated with ρ c = 1 12 as the value of c changes across the list of candidate models [75]. The independence of the different clusters of parameters has been considered in this paper, which allows the factorization of the prior distribution and Fisher information matrix in Equation (21). Additionally, we approximate the determinant of the Fisher information matrix using the complete likelihood, and we consider the uninformative Jeffrey’s prior for the distribution of each group of parameters. Hence, in our case, the MML optimization objective function is calculated as follows:
MessLens c 2 ( 1 + log ρ c ) + c 2 ( log N ) + 2 M d = 1 D log ω d + 2 d k = 1 M log p k + d = 1 D log ( 1 ω d ) log p ( X | Θ M )
Equation (22) is minimized with respect to several constraints [23], which are listed as follows: 0 < p k 1 , 0 ω d 1 , and j = 1 M p j = 1 . In the context of this model selection criterion, since we are estimating feature weights using the EM algorithm, Equations (23) and (24) are utilized alternatively to approximate the parameters p ^ k and ω ^ d , respectively, as follows:
p ^ k = max i = 1 N j = 1 M γ ( Z i j ) 2 D , 0 j = 1 M max i = 1 N γ ( Z i j ) 2 D , 0
ω ^ d = max i = 1 N j = 1 M ω d p ( X i d | θ j d ) ζ i , j , d γ ( Z i j ) 2 M , 0 T
T = max i = 1 N j = 1 M ω d p ( X i d | θ j d ) ζ i , j , d γ ( Z i j ) 2 M , 0 + max i = 1 N j = 1 M ( 1 ω d ) p ( X i d | β d ) ζ i , j , d γ ( Z i j ) 1 , 0
The Algorithm of Model Selection and Model Parameter Estimation
Algorithm 1 describes how to perform model selection and feature selection using the MML criterion and model parameter estimation using the EM algorithm.
Algorithm 1: Unsupervised FSBAGGMM
1:
While M < M m a x do
2:
     Initialize Θ M
  • K-means clustering results are used to initialize the parameters ( π 1 , , π M , μ 1 , , μ M , σ l 1 , , σ l M , σ r 1 , , σ r M , λ 1 , , λ M ) .
  • For each cluster k, each element of the parameter vector λ k is set to the value 2.
  • Initialize the background Gaussian distribution parameter set β using the following equations for all the dimensions, where d { 1 , , D } :
    η d = 1 N i = 1 N X i d
    δ d 2 = 1 N i = 1 N ( X i d η d ) 2
3:
     Implement the E-step.
  • For each cluster k, compute the bounded support region τ k = ( τ 1 , , τ D ) .
  • Evaluate Equation (12).
    • if  ω d = 0   Then  p ( X i d | θ k d ) = 0
    • if  ω d = 1   Then  p ( X i d | β d ) = 0
4:
     Implement the M-step using Equations (15) through (20), (23), and (24).
5:
     if  p ( X | Θ ) ι + 1 p ( X | Θ ) ι < ϵ  then
  • Calculate the message length using Equation (22).

4.3. Implementation with HPC

The advancements in computational methodologies have played a pivotal role in addressing the challenges of data processing, especially in the realm of smart meters. Given the magnitude and intricacy of the data generated by these meters, traditional computing methods often fall short. This necessitated the exploration and implementation of our algorithm via HPC.
Our choice of HPC was rooted in its inherent capability to expediently process large volumes of data. For the clustering task at hand, HPC provided the computational agility required to analyze vast datasets from smart meters swiftly. By leveraging the parallel processing capabilities of HPC, we could achieve a significant reduction in computation time, while ensuring the consistency and accuracy of our clustering results.
Edge cloud computing stands at the forefront of modern computational paradigms, emphasizing on-the-spot processing to facilitate real-time decision-making. With the integration of HPC in edge settings, we foresee several advantages.
  • Enhanced Speed and Efficiency: By employing HPC at the edge, data from smart meters can be processed locally, resulting in quicker analytics and response times. This is especially crucial for utility programs that require timely information, such as demand response and energy efficiency initiatives.
  • Scalability: As the deployment of smart meters expands, the amount of data to be processed will proportionally increase. HPC can readily handle this surge, ensuring that the system can scale without compromising on performance.
  • Real-Time Analytics for Utility Programs: HPC, coupled with edge cloud computing, can power real-time analytics. For instance, utility providers can swiftly analyze consumption patterns and roll out demand response strategies almost instantaneously. This not only enhances grid reliability but also aids in optimizing energy consumption and costs for consumers.

5. Experimental Results

In this section, we will validate the performance of the MML model selection criterion and the proposed FSBAGGMM using two synthetic and real-life smart meter datasets within the application of household energy consumption segmentation. The first real-life dataset was recorded by the Commission for Energy Regulation (CER) and made accessible for researchers by the Irish Social Science Data Archive (ISSDA) [4]. The dataset consists of smart meter data gathered from more than 6000 Irish energy consumers from 14 July 2009 to 31 December 2010. The energy consumption is recorded in kWh with an interval of half an hour. This dataset has two types of energy consumers: residential and small to medium enterprises. As stated earlier, we are interested in analyzing the energy consumption of residential energy consumers only. Therefore, 3639 Irish residential energy consumers remain for analysis after data cleaning. Each residential consumer is assigned six different tariffs (E, A, D, C, B, and W). The second real-life smart meter dataset consists of smart meter data collected from 5567 residential homes in London. The data were collected by the UK Power Networks led by the Low Carbon London Project between November 2011 and February 2014 [6]. The energy consumption is recorded in kWh with an interval of half an hour. After data cleaning, observations of 3891 household energy consumers within the year 2013 are used to analyze this experiment. The residential energy consumers in this dataset are subjected to two types of tariffs. The first type is the dynamic time of use (ToU), where the energy consumption prices vary as follows: high (67.20 pence/kWh), low (3.99 pence/kWh), or normal (11.76 pence/kWh). The second type is the standard (std), where the consumers pay a flat rate of 14.228 pence/kWh. Additionally, the energy consumers in this dataset belong to five different geo-demographic groups.
The application considered in this paper aims to segment energy consumers given their load curve. We use characteristic load profiles to find the optimal number of energy consumption clusters with similar consumption patterns and determine the cluster membership of every load curve given in the training dataset. Utility companies can use accurate energy-consumer-type identification to make correct decisions regarding the investments in load-shifting campaigns to prevent over- or under-dimensioning linked to the peak energy demand. Several performance evaluation metrics [64] are used in this paper. They are defined as follows.
DI [76]: Dunn’s index is a model performance evaluation metric that is calculated using the minimum ratio between the closest distance of two observations of different clusters and the largest distance between two observations in the same cluster. This index is maximized for the best clustering and it is defined as follows:
DI = min A M min B M , B A { ϕ ( A , B ) } max A M { Π ( A ) }
ϕ ( A , B ) = min X i A , Y j B d ( X i , Y j )
Π ( A ) = max X i , X j A d ( X i , X j )
where d denotes the distance or the similarity function, ϕ ( A , B ) denotes the minimum distance between two observations that each belong to either cluster A or B, and M denotes the set of clusters.
EoE [31]: The entropy of eigenvalues is an entropy-based clustering performance measure; it is obtained from the eigenvalue analysis of the correlation matrix calculated using raw smart meter data. The index is calculated using the correlation between representative time series of different clusters and the correlation between different time series within each cluster. The EoE index is calculated using the following equation:
EoE = S M B k K N k N S M w k
The SM similarity is a normalized average information measure; the larger it is, the greater the similarity. The term S M b represents the normalized entropy of eigenvalues obtained from the correlation matrix between different clusters, and S M w k represents the normalized entropy of eigenvalues obtained from the correlation matrix between time series in each cluster k. In an ideal clustering, EoE is a small value consisting of high similarity between time series within each cluster and low similarity between representative time series of different clusters.
S [77]: The silhouette score is a model evaluation measure that is concerned with calculating a score for each observation in the training dataset. The measure calculates the overall evaluation by computing the average score for all the dataset observations. The metric is maximized for better clustering and is defined in the following equation:
s ( x i ) = b ( x i ) a ( x i ) m a x { a ( x i ) , b ( x i ) }
where a ( x i ) represents the average dissimilarity of the data point x i to all the other data points within the same cluster. b ( x i ) represents the minimum average dissimilarity of data point x i to data points existing in a cluster different from the data point’s cluster.
CH [78]: The Calinski–Harabasz index is a model performance evaluation index; the measure calculates the ratio between the inter-cluster variance and the intra-cluster variance. This measure is maximized for better clustering and is defined as follows:
CH = N K K 1 k = 1 K N k d ( c k , c ¯ ) k = 1 K i = 1 N k d ( X i , c k )
where N k is the number of observations predicted to belong to cluster k, c k denotes the centroid of class k, c ¯ denotes the global centroid of all the clusters, and d denotes the distance or the similarity function.
DB [79]: The Davies–Bouldin index is a model performance evaluation measure; it calculates the ratio of intra-cluster distances to inter-cluster distances for each possible pair of clusters. The maximum ratio calculated for each pair of clusters is considered in a summation. The summation result is divided by the total number of clusters to obtain the metric’s value. This measure is minimized for better clustering, and it is defined as follows:
DB = 1 k A M max B M , B A O ( A ) + O ( B ) d ( c A , c B )
O ( A ) = 1 ϱ ( A ) X i A d ( X i , c A )
where ϱ ( A ) denotes the cardinality of cluster A, k denotes the number of components enforced by the mixture model, M denotes the set of clusters, c A denotes the centroid of class A, and d denotes the distance or the similarity function. M has k elements.
GOF [80]: The goodness of fit statistic value measures the model’s fitting accuracy and it is calculated as follows:
GOF = i = 1 N ( Υ ( X i ) Ω ( X i ) ) 2 Ω ( X i )
where Υ ( X i ) and Ω ( X i ) represent the empirical and the expected frequencies of the observation X i , respectively. The indices ACC, TPR, PPV, TNR, NPV, FPR, FNR, and FDR represent the average accuracy, average true positive rate, positive predictive value, true negative rate, negative predictive value, false positive rate, false negative rate, and false discovery rate, respectively. They are defined as follows:
TPR = 1 M k = 1 M TP k TP k + FN k
TNR = 1 M k = 1 M TN k TN k + FP k
PPV = 1 M k = 1 M TP k TP k + FP k
NPV = 1 M k = 1 M TN k TN k + FN k
FPR = 1 M k = 1 M FP k FP k + TN k
FNR = 1 M k = 1 M FN k TP k + FN k
FDR = 1 M k = 1 M FP k TP k + FP k
ACC = 1 M k = 1 M TP k + TN k TP k + FP k + FN k + TN k
where TP k , FP k , TN k , and FN k denote the number of true positives, false positives, true negatives, and false negatives, respectively, for the cluster k. In order to compute the metrics explained in Equations (37)–(44), cluster k labels are considered a positive class and all the remaining cluster labels are considered a negative class. MCC represents the Matthews correlation coefficient evaluation metric [81].
The AIC and BIC are probabilistic model selection methods [82] that attempt to select the model with the best performance while taking into consideration its complexity (by adding a complexity-related penalty). Unlike probabilistic model selection criteria, performance metrics select models with no regard to their complexity. The distinct probabilistic model selection criteria used in this paper originate from different fields of study. The AIC is derived from the frequentist framework, while the BIC is derived from Bayesian probability and inference. Compared to the BIC, the AIC emphasizes the model performance and penalizes complex models less, making it prone to selecting overfitted models. In comparison to the AIC, the BIC attempts to penalize candidate models more for their complexity. The AIC and BIC model selection criteria statistics for each candidate model are computed as follows:
B I C = 2 log ( L ( Θ ) ) + κ log ( N )
A I C = 2 N log ( L ( Θ ) ) + 2 κ N
where L ( Θ ) is the likelihood function estimate given a set of parameters Θ , κ represents the number of free parameters, and N represents the number of observations. As N approaches infinity, the BIC criterion is more likely to select the candidate model with the true number of intrinsic clusters. The candidate model with the lowest AIC and BIC is selected for both model selection criteria.
In the upcoming sections, the performance of the proposed model is compared to specific mixture models such as the BAGGMM, the AGGMM, and the FSAGGMM. Model selection using the proposed model is performed using the MML model selection criterion and compared against specific model selection methods such as the BIC and AIC, and model selection methods using performance measures, such as Dunn’s index (DI) and the entropy of eigenvalues (EoE).

5.1. Synthetic Data

As a first stage, synthetic datasets are used to validate the proposed mixture model and its model selection method. We propose using a 49-dimensional dataset, which imitates a real-life smart meter dataset by representing each energy consumer with a load curve. In order to generate the synthetic datasets used in this paper, the following steps were followed.
  • For each energy consumer in the real-life dataset, only the first 49 smart meter observations are considered.
  • The Gaussian mixture model is used to cluster the data into a specific number of clusters. The mean of each cluster is considered a consumption profile.
  • Each consumption profile inferred from the previous step is summed with instances generated by Gaussian white noise using five different sets of parameters to form the observations of the synthetic dataset.
In other words, the origin of each cluster of observations within the synthetic datasets used in this paper is an actual energy consumption profile concluded from a real dataset.
The data-generating process delineated above provides a systematic approach to crafting synthetic datasets with asymmetric class distributions and varied shapes. By grounding the data in real consumption profiles and subsequently introducing variations via Gaussian white noise, the process ensures a rich diversity of data shapes. This diversity serves as a rigorous testing ground to evaluate the flexibility and robustness of the proposed mixture model, effectively challenging its capability to adapt and accurately represent varied data structures.
The first dataset consists of five clusters. The five real-life consumption profiles used to generate the first dataset are demonstrated in Figure 2a. The count of the observations generated for each energy consumption profile using the distinct Gaussian white noise parameters is shown in Table 2. The clustering results of our proposed model are evaluated using several performance measures and compared against the clustering performance of specific mixture models, as shown in Table 3 and Table 4. As an illustrative example of the data generation process, 378 observations of the first dataset are generated by summing the white noise vector generated using the parameter set ( μ = 0.001 ; σ = 0.2 ) of the multivariate Gaussian white noise with the vector of “Consumption Profile 1”.
Our model selection approach successfully infers the correct number of components within this dataset, as demonstrated in Table 5. MML outperforms specific model selection methods using the clustering results obtained from each instance of our proposed model.
Figure 3a demonstrates the maximum log-likelihood achieved by clustering the data using the proposed model in comparison with specific mixture models. The proposed model achieves the best fit of the training data by achieving the best performance according to all the performance metrics used in this experiment and by reaching the highest log-likelihood.
The second dataset consists of eight clusters. The eight real-life consumption profiles used to generate this dataset are demonstrated in Figure 2b. Our model selection approach successfully infers the correct number of components within this dataset, as demonstrated in Table 6. The count of the observations generated for each energy consumption profile using the distinct Gaussian white noise parameters is shown in Table 7. MML chooses the proposed model’s instance with a component count equal to the ground truth, outperforming specific model selection methods used in this comparison. The proposed model fits the data better than all the mixture models used in the comparison by achieving the highest maximum log-likelihood, as demonstrated in Figure 3b. According to all the performance metrics used in this experiment, the proposed model also outperforms the mixture models selected for the comparison, as shown in Table 8 and Table 9.

5.2. Real-Life Smart Meter Data

5.2.1. The Commission for Energy Regulation Smart Meter Data

In this section, we investigate the performance of our proposed model using the first real-life smart meter dataset. As mentioned earlier, the dataset that we consider has smart meter observations from 3639 Irish energy consumers. Each consumer has 25,728 electricity usage readings that are recorded in kilowatt-hours. In order to summarize and preserve the information within the numerous features representing each energy consumer, PCA is used for feature extraction in this experiment. Several datasets with a different number of features are considered within the range between 50 and 250. Due to the low reconstruction error, the dataset with 250 features is favoured for this experiment.
We used the dataset as an input to three different instances of our proposed model. Each instance had a different number of mixture components within the range M = [ 2 , 4 ] . The model selection algorithm concluded that the minimum value calculated using its objective function was obtained while using the model instance with three components, as shown in Figure 4a. Table 10 demonstrates the optimal number of clusters concluded by each model selection criterion used in comparison with MML. In addition to the fact that our derived model selection criterion infers the correct number of clusters in solid experiments using synthetic data, the AIC and BIC also agree that the true number of clusters is three in this experiment.
Figure 4b demonstrates the log-likelihood trail for each mixture model used in the comparison within this experiment. As observed, the proposed model converged to the highest log-likelihood, indicating a better fit to the training dataset. The clustering evaluation of the proposed model for the concluded optimal number of clusters is demonstrated in Table 11 in comparison with specific mixture models. As demonstrated, our proposed model achieves the best clustering performance according to all the evaluation measures used in the comparison.
As mentioned earlier, we determined the true number of clusters using MML and achieved the best clustering result using our proposed mixture model. Since this is an implementation of a real-life application, it is necessary to analyze the resulting clusters to understand further the energy consumption patterns of each consumption trend discovered. Figure 5a demonstrates the average power demand of all the energy consumers without clustering. Comparatively, we demonstrate the average power demand of each energy consumer cluster in Figure 5b. For all the time intervals available in the dataset, as observed, the responsibility of each energy consumption pattern to the overall average power demand can be determined. The proposed model can determine the consumer’s contribution to each consumption profile and which the consumer is mostly following. Table 12 demonstrates the ratio of the count of energy consumers in each cluster to the total count of energy consumers in the dataset; the table also demonstrates the consumption responsibility of each consumer cluster to the total average energy consumption in the year 2010. Additionally, the real-life dataset that we use in this experiment provides the tariff assigned for each energy consumer. We have discovered that the tariff types are distributed almost identically across the resulting clusters, as shown in Figure 6, which indicates that the tariff type does not influence the consumer’s electrical usage pattern.

5.2.2. The UK Power Networks Smart Meter Data

In this section, we validate the performance of our proposed model using the second real-life smart meter dataset. As mentioned earlier, the dataset that we consider in this experiment has smart meter observations from 3891 household energy consumers that are located in London. Each consumer has 17,520 electricity usage readings that are recorded in kilowatt-hours. In order to summarize the information included in the load curve of each energy consumer, we have extracted nine features. Following [32], seven features are extracted after the definition of four key time periods and they are denoted by t { 1 , 2 , 3 , 4 } . The overnight time period ( t = 1 ) is defined between 10:30 p.m. and 6:30 a.m., the breakfast time period ( t = 2 ) is defined between 6:30 a.m. and 9:00 a.m., the daytime period ( t = 3 ) is defined between 9:00 a.m. and 3:30 p.m., and the evening time period ( t = 4 ) is defined between 3:30 p.m. and 10:30 p.m. Based on the four previously explained prominent time periods, seven features are extracted from the smart meter data to summarize the representation of energy consumers, and they are calculated as follows.
  • RAP t denotes the relative average power for time period (t) over the entire year; it is defined as follows:
    RAP t = AP t DAP , t = 1 , 2 , 3 , 4
  • the mean STD denotes the mean relative standard deviation of the average power used over the entire year; it is defined as follows:
    Mean STD = 1 4 t = 1 4 σ t AP t
  • The seasonal score is defined as follows:
    Seasonal Score = t = 1 4 | A P t W A P t S | A P t
  • The weekend vs. weekday difference score (WD-WE diff. score) is calculated as follows:
WD WE diff . Score = i = 1 4 | AP t WD AP t WE | AP t
where AP t , and σ t represent the average power used by the specific consumer and its corresponding standard deviation in the time period (t), respectively, over all the available smart meter data. DAP represents the average daily power used by the specific consumer throughout the available smart meter data. A P t W and A P t S represent the average power used by the specific consumer in the time period (t) throughout winter and summer, respectively. AP t WD , and AP t WE represent the average power used by the specific consumer in the time period (t) throughout the weekdays and weekends, respectively, for the available data. Finally, the eighth and the ninth features represent the consumer’s tariff and geo-demographic group, respectively.
We have determined the optimal number of clusters for our proposed model using the MML model selection criterion, similarly to our previous experiments. Among five candidate FSBAGGMM models of mixture components within the range [2, 6], the model instance with four components achieved the minimum message length.
Most of the model selection methods used in the comparison demonstrated in Table 13 agree on the optimal number of mixture components. Therefore, the data were clustered into four clusters using our proposed model, and the clustering performance evaluation was compared against specific mixture models. Table 14 demonstrates how our proposed mixture model has been able to outperform the different mixture models used in the comparison using six different performance metrics.
As shown in Figure 7b, the categorical feature representing the tariff for each energy consumer has an almost identical distribution across the clusters obtained using our proposed mixture model, having little to no influence on the energy consumption behaviour. Nevertheless, as demonstrated by the CH score in Table 14, our proposed model has achieved clusters with relatively small intra-cluster (within clusters) variance and relatively large inter-cluster (between clusters) variance. Additionally, the minimum number of members within the clusters achieved using the FSBAGGMM is 225 energy consumers, as demonstrated in Figure 7a. Additionally, Table 15 demonstrates the average values of several features for the inferred household energy consumer clusters.
Since the smart meter data have been modelled successfully, the proposed model is capable of identifying energy consumer clusters that are suitable for demand reduction initiatives within several utility programs [2]. As an example, Table 15 demonstrates that the first cluster has a relatively high evening RAP with a relatively low mean STD, seasonal score, and WD-WE difference score. The power demand of energy consumers exhibiting energy consumption patterns similar to the first cluster could be lowered by implementing storage devices. The third and fourth clusters’ energy consumption patterns exhibit relatively low variability in demand, as represented by the mean STD and WD-WE difference score, while exhibiting a relatively high seasonal difference in power demand, as represented by the seasonal score. Such households could be offered non-electric or more efficient heating systems to reduce the winter demand.

6. Discussion

In this paper, we have presented an expectation-maximization algorithm within the MML criterion to optimize the parameters of the bounded asymmetric generalized Gaussian mixture model and to find the optimal number of consumption profiles and the optimal subset of features simultaneously. Our approach assumes that the data arise from a mixture of bounded asymmetric generalized Gaussian distributions. The final results demonstrate that the load curve of an individual energy consumer shows a probabilistic association with each class, indicating which pattern of electricity use is more or less likely to be used within a household. Therefore, it is possible to categorize households and how they consume energy using our proposed model.
Prior works in household energy consumption segmentation unrealistically approach model selection and feature selection as independent problems. Our approach successfully achieves the discovery of the true number of energy consumption profiles and the determination of the optimal set of data attributes to be used for modelling in our proposed mixture model in a single optimization process and avoids running the EM algorithm many times.
Clustering synthetically generated smart meter data with a ground truth cluster size, our proposed algorithm has outperformed most of the existing model selection approaches. In the same experiment, the proposed model correctly models the first and the second synthetic smart meter data with high accuracy of 95.569% and 91.856%, respectively. Similarly, our algorithm has also determined the optimal number of clusters in both datasets in experiments involving real-life data, and the proposed model outperforms all the mixture models used in the comparison, as demonstrated by all the utilized performance metrics. Thus, the superiority of the proposed algorithm in modelling smart meter data with different feature extraction methods over all the state-of-the-art clustering algorithms used in the comparison is proven.
Privacy and security concerns loom large in the realm of smart meter data analytics. Fortunately, the datasets employed in our research have been thoughtfully curated, with a paramount emphasis on safeguarding the privacy of individuals whose households are equipped with smart meters. These datasets meticulously exclude any information that might compromise the privacy of the participants while providing valuable insights for research. We have underscored in our research paper, particularly in the Results section, that the conventional categorization, carried out prior to any consumption data observation, is fundamentally ineffective. Respecting individuals’ privacy is not only an ethical imperative but also a fundamental human right. Remarkably, our proposed mixture model navigates this privacy-centric landscape adeptly. It uncovers the underlying data distribution and identifies energy consumption patterns without the need for additional, potentially intrusive information. This privacy-preserving approach aligns with the broader scientific quest for generalization and effectiveness in solutions that refrain from privacy invasion. Furthermore, our experiments with real-life datasets, which encompassed features such as tariff and geo-demographic groups, yielded intriguing results. These attributes, often considered vital, were deemed unimportant by our meticulous feature selection approach. This underscores our commitment to privacy and our ability to derive meaningful insights without resorting to invasive practices.
Finally, our implementation underscores a promising synergy between HPC and edge cloud computing, especially in the realm of smart meter data processing. As we progress towards a more interconnected and data-centric world, the amalgamation of these technologies will prove indispensable in sculpting the future of energy management and utility programs.

7. Conclusions

Our approach to analyzing real-life smart meter data is effective in determining households that are suitable for demand reduction initiatives such as DR and EE, thus providing the opportunity for utility companies to adopt environmentally friendly and cost-effective technologies.
The application addressed in this paper is well suited for an unsupervised approach, especially given the absence of ground truth labels. However, many applications would benefit from supervised or semi-supervised machine learning solutions. A limitation of the current learning framework presented in this paper is its inability to leverage ground truth labels. Recognizing this as a crucial area of improvement, future work could involve proposing a learning method for the mixture model that incorporates these labels to optimize the model parameters.

Author Contributions

Conceptualization, H.A.-B., M.A. (Muhammad Azam), M.A. (Manar Amayri) and N.B.; Data curation, H.A.-B.; Formal analysis, H.A.-B.; Investigation, H.A.-B.; Methodology, H.A.-B.; Resources, M.A. (Manar Amayri) and N.B.; Software, H.A.-B.; Supervision, M.A. (Manar Amayri) and N.B.; Validation, H.A.-B.; Visualization, H.A.-B.; Writing—original draft, H.A.-B.; Writing—review and editing, H.A.-B., M.A. (Muhammad Azam), M.A. (Manar Amayri) and N.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no funding.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to ethical, legal or privacy issues.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Important Partial Derivatives

ln Ψ ( X i d | θ k d ) μ k d = ln g 1 ( X i d | θ k d ) μ k d x < μ k d ln g 2 ( X i d | θ k d ) μ k d x μ k d = A ( λ k d ) λ k d ( μ k d X i d ) λ k d 1 σ l k d λ k d x < μ k d A ( λ k d ) λ k d ( X i d μ k d ) λ k d 1 σ r k d λ k d x μ k d
L ( X , Θ M , Z , φ ) μ k d = i = 1 N ω d p ( x i d | θ k d ) ζ i k d p ( k | X i , Θ M ) × ln g 1 ( X i d | θ k d ) μ k d + k g 1 ( X i d | θ k d ) ln g 1 ( X i d | θ k d ) μ k d d u k g 1 ( X i d | θ k d ) d u x < μ k d ln g 2 ( X i d | θ k d ) μ k d + k g 2 ( X i d | θ k d ) ln g 2 ( X i d | θ k d ) μ k d d u k g 2 ( X i d | θ k d ) d u x μ k d
ln Ψ ( X i d | θ k d ) σ l k d = ln g 1 ( X i d | θ k d ) σ l k d x < μ k d ln g 2 ( X i d | θ k d ) σ l k d x μ k d = A ( λ k d ) λ k d ( μ k d X i d ) λ k d σ l k d λ k d + 1 1 σ l k d + σ r k d x < μ k d 1 σ l k d + σ r k d x μ k d
L ( X , Θ M , Z , φ ) σ l k d = i = 1 N ω d p ( x i d | θ k d ) ζ i k d p ( k | X i , Θ M ) × ln g 1 ( X i d | θ k d ) σ l k d + k g 1 ( X i d | θ k d ) ln g 1 ( X i d | θ k d ) σ l k d d u k g 1 ( X i d | θ k d ) d u x < μ k d ln g 2 ( X i d | θ k d ) σ l k d + k g 2 ( X i d | θ k d ) ln g 2 ( X i d | θ k d ) σ l k d d u k g 2 ( X i d | θ k d ) d u x μ k d
ln Ψ ( X i d | θ k d ) σ r k d = ln g 1 ( X i d | θ k d ) σ r k d x < μ k d ln g 2 ( X i d | θ k d ) σ r k d x μ k d = 1 σ l k d + σ r k d x < μ k d A ( λ k d ) λ k d σ r k d ( X i d μ k d ) λ k d σ r k d λ k d 1 σ l k d + σ r k d x μ k d
L ( X , Θ M , Z , φ ) σ r k d = i = 1 N ω d p ( x i d | θ k d ) ζ i k d p ( k | X i , Θ M ) × ln g 1 ( X i d | θ k d ) σ r k d + k g 1 ( X i d | θ k d ) ln g 1 ( X i d | θ k d ) σ r k d d u k g 1 ( X i d | θ k d ) d u x < μ k d ln g 2 ( X i d | θ k d ) σ r k d + k g 2 ( X i d | θ k d ) ln g 2 ( X i d | θ k d ) σ r k d d u k g 2 ( X i d | θ k d ) d u x μ k d
ln Ψ ( X i d | θ k d ) λ k d = ln g 1 ( X i d | θ k d ) λ k d x < μ k d ln g 2 ( X i d | θ k d ) λ k d x μ k d = 1 λ k d + 3 ( ψ ( 1 / λ k d ) ψ ( 3 / λ k d ) ) 2 λ k d 2 μ k d X i d σ l k d λ k d A ( λ k d ) × [ 1 2 ln Γ ( 3 / λ k d ) Γ ( 1 / λ k d ) + ( ψ ( 1 / λ k d ) 3 ψ ( 3 / λ k d ) ) 2 λ k d + ln μ k d X i d σ l k d ] x < μ k d 1 λ k d + 3 ( ψ ( 1 / λ k d ) ψ ( 3 / λ k d ) ) 2 λ k d 2 X i d μ k d σ r k d λ k d A ( λ k d ) × [ 1 2 ln Γ ( 3 / λ k d ) Γ ( 1 / λ k d ) + ( ψ ( 1 / λ k d ) 3 ψ ( 3 / λ k d ) ) 2 λ k d + ln X i d μ k d σ r k d ] x μ k d
L ( X , Θ M , Z , φ ) λ k d = i = 1 N ω d p ( x i d | θ k d ) ζ i k d p ( k | X i , Θ M ) × ln g 1 ( X i d | θ k d ) λ k d + k g 1 ( X i d | θ k d ) ln g 1 ( X i d | θ k d ) λ k d d u k g 1 ( X i d | θ k d ) d u x < μ k d ln g 2 ( X i d | θ k d ) λ k d + k g 2 ( X i d | θ k d ) ln g 2 ( X i d | θ k d ) λ k d d u k g 2 ( X i d | θ k d ) d u x μ k d
2 L ( X , Θ M , Z , φ ) μ k d 2 = i = 1 N ω d p ( x i d | θ k d ) ζ i k d p ( k | X i , Θ M ) × 2 ln g 1 ( X i d | θ k d ) μ k d 2 + k g 1 ( X i d | θ k d ) ln g 1 ( X i d | θ k d ) μ k d 2 + 2 ln g 1 ( X i d | θ k d ) μ k d 2 d u k g 1 ( X i d | θ k d ) d u k g 1 ( X i d | θ k d ) ln g 1 ( X i d | θ k d ) μ k d d u 2 k g 1 ( X i d | θ k d ) d u 2 x < μ k d 2 ln g 2 ( X i d | θ k d ) μ k d 2 + k g 2 ( X i d | θ k d ) ln g 2 ( X i d | θ k d ) μ k d 2 + 2 ln g 2 ( X i d | θ k d ) μ k d 2 d u k g 2 ( X i d | θ k d ) d u k g 2 ( X i d | θ k d ) ln g 2 ( X i d | θ k d ) μ k d d u 2 k g 2 ( X i d | θ k d ) d u 2 x μ k d
2 ln Ψ ( X i d | θ k d ) μ k d 2 = 2 ln g 1 ( X i d | θ k d ) μ k d 2 x < μ k d 2 ln g 2 ( X i d | θ k d ) μ k d 2 x μ k d = A ( λ k d ) λ k d ( λ k d 1 ) ( μ k d X i d ) λ k d 2 σ k d λ k d x < μ k d A ( λ k d ) λ k d ( λ k d 1 ) ( X i d μ k d ) λ k d 2 σ k d λ k d x μ k d
2 L ( X , Θ M , Z , φ ) σ l k d 2 = i = 1 N ω d p ( x i d | θ k d ) ζ i k d p ( k | X i , Θ M ) × 2 ln g 1 ( X i d | θ k d ) σ l k d 2 + k g 1 ( X i d | θ k d ) ln g 1 ( X i d | θ k d ) σ l k d 2 + 2 ln g 1 ( X i d | θ k d ) σ l k d 2 d u k g 1 ( X i d | θ k d ) d u k g 1 ( X i d | θ k d ) ln g 1 ( X i d | θ k d ) σ l k d d u 2 k g 1 ( X i d | θ k d ) d u 2 x < μ k d 2 ln g 2 ( X i d | θ k d ) σ l k d 2 + k g 2 ( X i d | θ k d ) ln g 2 ( X i d | θ k d ) σ l k d 2 + 2 ln g 2 ( X i d | θ k d ) σ l k d 2 d u k g 2 ( X i d | θ k d ) d u k g 2 ( X i d | θ k d ) ln g 2 ( X i d | θ k d ) σ l k d d u 2 k g 2 ( X i d | θ k d ) d u 2 x μ k d
2 ln Ψ ( X i d | θ k d ) σ l k d 2 = 2 ln g 1 ( X i d | θ k d ) σ l k d 2 x < μ k d 2 ln g 2 ( X i d | θ k d ) σ l k d 2 x μ k d = ( λ k d + 1 ) A ( λ k d ) λ k d ( μ k d X i d ) λ k d σ l k d λ k d + 2 + 1 ( σ l k d + σ r k d ) 2 x < μ k d 1 ( σ l k d + σ r k d ) 2 x μ k d
2 L ( X , Θ M , Z , φ ) σ r k d 2 = i = 1 N ω d p ( x i d | θ k d ) ζ i k d p ( k | X i , Θ M ) × 2 ln g 1 ( X i d | θ k d ) σ r k d 2 + k g 1 ( X i d | θ k d ) ln g 1 ( X i d | θ k d ) σ r k d 2 + 2 ln g 1 ( X i d | θ k d ) σ r k d 2 d u k g 1 ( X i d | θ k d ) d u k g 1 ( X i d | θ k d ) ln g 1 ( X i d | θ k d ) σ r k d d u 2 k g 1 ( X i d | θ k d ) d u 2 x < μ k d 2 ln g 2 ( X i d | θ k d ) σ r k d 2 + k g 2 ( X i d | θ k d ) ln g 2 ( X i d | θ k d ) σ r k d 2 + 2 ln g 2 ( X i d | θ k d ) σ r k d 2 d u k g 2 ( X i d | θ k d ) d u k g 2 ( X i d | θ k d ) ln g 2 ( X i d | θ k d ) σ r k d d u 2 k g 2 ( X i d | θ k d ) d u 2 x μ k d
2 ln Ψ ( X i d | θ k d ) σ r k d 2 = 2 ln g 1 ( X i d | θ k d ) σ r k d 2 x < μ k d 2 ln g 2 ( X i d | θ k d ) σ r k d 2 x μ k d = 1 ( σ l k d + σ r k d ) 2 x < μ k d ( λ k d + 1 ) A ( λ k d ) λ k d ( X i d μ k d ) λ k d σ r k d λ k d + 2 + 1 ( σ l k d + σ r k d ) 2 x μ k d
2 L ( X , Θ M , Z , φ ) λ k d 2 = i = 1 N ω d p ( x i d | θ k d ) ζ i k d p ( k | X i , Θ M ) × 2 ln g 1 ( X i d | θ k d ) λ k d 2 + k g 1 ( X i d | θ k d ) ln g 1 ( X i d | θ k d ) λ k d 2 + 2 ln g 1 ( X i d | θ k d ) λ k d 2 d u k g 1 ( X i d | θ k d ) d u k g 1 ( X i d | θ k d ) ln g 1 ( X i d | θ k d ) λ k d d u 2 k g 1 ( X i d | θ k d ) d u 2 x < μ k d 2 ln g 2 ( X i d | θ k d ) λ k d 2 + k g 2 ( X i d | θ k d ) ln g 2 ( X i d | θ k d ) λ k d 2 + 2 ln g 2 ( X i d | θ k d ) λ k d 2 d u k g 2 ( X i d | θ k d ) d u k g 2 ( X i d | θ k d ) ln g 2 ( X i d | θ k d ) λ k d d u 2 k g 2 ( X i d | θ k d ) d u 2 x μ k d
2 ln Ψ ( X i d | θ k d ) λ k d 2 = 2 ln g 1 ( X i d | θ k d ) λ k d 2 x < μ k d 2 ln g 2 ( X i d | θ k d ) λ k d 2 x μ k d = 1 λ k d 2 + 3 ( ψ ( 3 / λ k d ) ψ ( 1 / λ k d ) ) 2 λ k d 4 3 ( ψ ( 1 / λ k d ) ψ ( 3 / λ k d ) ) λ k d 3 μ k d X i d σ l k d λ k d A ( λ k d ) × [ ( 9 ψ ( 3 / λ k d ) ψ ( 1 / λ k d ) ) 2 λ k d 3 + [ 1 2 ln Γ ( 3 / λ k d ) Γ ( 1 / λ k d ) + ( ψ ( 1 / λ k d ) 3 ψ ( 3 / λ k d ) ) 2 λ k d + ln μ k d X i d σ l k d ] 2 ] x < μ k d 1 λ k d 2 + 3 ( ψ ( 3 / λ k d ) ψ ( 1 / λ k d ) ) 2 λ k d 4 3 ( ψ ( 1 / λ k d ) ψ ( 3 / λ k d ) ) λ k d 3 X i d μ k d σ r k d λ k d A ( λ k d ) × [ ( 9 ψ ( 3 / λ k d ) ψ ( 1 / λ k d ) ) 2 λ k d 3 + [ 1 2 ln Γ ( 3 / λ k d ) Γ ( 1 / λ k d ) + ( ψ ( 1 / λ k d ) 3 ψ ( 3 / λ k d ) ) 2 λ k d + ln X i d μ k d σ r k d ] 2 ] x μ k d

References

  1. Lloret, J.; Tomas, J.; Canovas, A.; Parra, L. An integrated IoT architecture for smart metering. IEEE Commun. Mag. 2016, 54, 50–57. [Google Scholar] [CrossRef]
  2. Kwac, J.; Flora, J.; Rajagopal, R. Household energy consumption segmentation using hourly data. IEEE Trans. Smart Grid 2014, 5, 420–430. [Google Scholar] [CrossRef]
  3. Haben, S.; Ward, J.; Greetham, D.; Singleton, C.; Grindrod, P.A. New error measure for forecasts of household-level, high resolution electrical energy consumption. Int. J. Forecast. 2014, 30, 246–256. [Google Scholar] [CrossRef]
  4. CER. CER Smart Metering Project - Electricity Customer Behaviour Trial, 2009–2010 [dataset]. 2012. Available online: https://www.ucd.ie/issda/data/commissionforenergyregulationcer/ (accessed on 5 April 2023).
  5. Cao, H.; Beckel, C.; Staake, T. Are domestic load profiles stable over time? An attempt to identify target households for demand side management campaigns. In Proceedings of the IECON 2013-39th Annual Conference of The IEEE Industrial Electronics Society, Vienna, Austria, 10–13 November 2013; pp. 4733–4738. [Google Scholar]
  6. UK Power Networks. SmartMeter Energy Consumption Data in London Households, 2011–2014 [dataset]. 2013. Available online: https://data.london.gov.uk/dataset/smartmeter-energy-use-data-in-london-households (accessed on 28 September 2023).
  7. Massachusetts (Amherst), (UMass Smart* Dataset—Microgrid Dataset, 2013 Release [dataset]. 2013. Available online: https://traces.cs.umass.edu/index.php/Smart/Smart (accessed on 28 September 2023).
  8. Alahakoon, D.; Yu, X. Smart electricity meter data intelligence for future energy systems: A survey. IEEE Trans. Ind. Inform. 2015, 12, 425–436. [Google Scholar] [CrossRef]
  9. Al Khafaf, N.; Jalili, M.; Sokolowski, P. Demand Response Planning Tool using Markov Decision Process. In Proceedings of the 2018 IEEE 16th International Conference on Industrial Informatics (INDIN), Porto, Portugal, 18–20 July 2018; pp. 484–489. [Google Scholar]
  10. Shahzadeh, A.; Khosravi, A.; Nahavandi, S. Improving load forecast accuracy by clustering consumers using smart meter data. In Proceedings of the 2015 International Joint Conference On Neural Networks (IJCNN), Killarney, Ireland, 12–17 July 2015; pp. 1–7. [Google Scholar]
  11. Chicco, G.; Napoli, R.; Postolache, P.; Scutariu, M.; Toader, C. Customer characterization options for improving the tariff offer. IEEE Trans. Power Syst. 2003, 18, 381–387. [Google Scholar] [CrossRef]
  12. Stephenson, P.; Lungu, I.; Paun, M.; Silvas, I.; Tupu, G. Tariff development for consumer groups in internal European electricity markets. In Proceedings of the 16th International Conference and Exhibition on Electricity Distribution, Amsterdam, The Netherlands, 18–21 June 2001; Part 1: Contributions. CIRED. (IEE Conf. Publ No. 482). Volume 5, p. 5. [Google Scholar]
  13. Chen, C.; Kang, M.; Hwang, J.; Huang, C. Synthesis of power system load profiles by class load study. Int. J. Electr. Power Energy Syst. 2000, 22, 325–330. [Google Scholar] [CrossRef]
  14. Meignen, S.; Meignen, H. On the modeling of small sample distributions with generalized Gaussian density in a maximum likelihood framework. IEEE Trans. Image Process. 2006, 15, 1647–1652. [Google Scholar] [CrossRef] [PubMed]
  15. Wang, D.; Xie, W.; Pei, J.; Lu, Z. Moving area detection based on estimation of static background. J. Inform. Comput. Sci. 2005, 2, 129–134. [Google Scholar]
  16. Palacios, M.; Steel, M. Non-gaussian bayesian geostatistical modeling. J. Am. Stat. Assoc. 2006, 101, 604–618. [Google Scholar] [CrossRef]
  17. Hedelin, P.; Skoglund, J. Vector quantization based on Gaussian mixture models. IEEE Trans. Speech Audio Process. 2000, 8, 385–401. [Google Scholar] [CrossRef]
  18. Nguyen, T.; Wu, Q.; Zhang, H. Bounded generalized Gaussian mixture model. Pattern Recognit. 2014, 47, 3132–3142. [Google Scholar] [CrossRef]
  19. Azam, M.; Bouguila, N. Bounded generalized gaussian mixture model with ica. Neural Process. Lett. 2019, 49, 1299–1320. [Google Scholar] [CrossRef]
  20. Lindblom, J.; Samuelsson, J. Bounded support Gaussian mixture modeling of speech spectra. IEEE Trans. Speech Audio Process. 2003, 11, 88–99. [Google Scholar] [CrossRef]
  21. Azam, M.; Bouguila, N. Multivariate bounded support asymmetric generalized Gaussian mixture model with model selection using minimum message length. Expert Syst. Appl. 2022, 204, 117516. [Google Scholar] [CrossRef]
  22. Raudys, S.; Jain, A.; Small, O. Sample size effects in statistical pattern recognition: Recommendations for practitioners. IEEE Trans. Pattern Anal. Mach. Intell. 1991, 13, 252–264. [Google Scholar] [CrossRef]
  23. Law, M.; Figueiredo, M.; Jain, A. Simultaneous feature selection and clustering using mixture models. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 1154–1166. [Google Scholar] [CrossRef] [PubMed]
  24. Pudil, P.; Novovičová, J.; Kittler, J. Floating search methods in feature selection. Pattern Recognit. Lett. 1994, 15, 1119–1125. [Google Scholar] [CrossRef]
  25. Kohavi, R.; John, G. Wrappers for feature subset selection. Artif. Intell. 1997, 97, 273–324. [Google Scholar] [CrossRef]
  26. Caruana, R.; Freitag, D. Greedy attribute selection. In Machine Learning Proceedings 1994; Morgan Kaufmann: Cambridge, MA, USA, 1994; pp. 28–36. [Google Scholar]
  27. Concordia University. High-Performance Computing Facility: Speed [Computing Resource]. 2018. Available online: https://www.concordia.ca/ginacody/aits/speed.html (accessed on 28 September 2023).
  28. Rafati, A.; Shaker, H.; Ghahghahzadeh, S. Fault Detection and Efficiency Assessment for HVAC Systems Using Non-Intrusive Load Monitoring: A Review. Energies 2022, 15, 341. [Google Scholar] [CrossRef]
  29. Rodríguez, M.; Cortés, A.; González Alonso, I.; Zalama Casanova, E. Using the Big Data generated by the Smart Home to improve energy efficiency management. Energy Effic. 2016, 9, 249–260. [Google Scholar] [CrossRef]
  30. Liu, X.; Nielsen, P.A. hybrid ICT-solution for smart meter data analytics. Energy 2016, 115, 1710–1722. [Google Scholar] [CrossRef]
  31. Al Khafaf, N.; Jalili, M.; Sokolowski, P. A novel clustering index to find optimal clusters size with application to segmentation of energy consumers. IEEE Trans. Ind. Inform. 2020, 17, 346–355. [Google Scholar] [CrossRef]
  32. Haben, S.; Singleton, C.; Grindrod, P. Analysis and clustering of residential customers energy behavioral demand using smart meter data. IEEE Trans. Smart Grid 2015, 7, 136–144. [Google Scholar] [CrossRef]
  33. Albert, A.; Rajagopal, R. Smart meter driven segmentation: What your consumption says about you. IEEE Trans. Power Syst. 2013, 28, 4019–4030. [Google Scholar] [CrossRef]
  34. Erdem, T.; Eken, S. Layer-Wise Relevance Propagation for Smart-Grid Stability Prediction. In Mediterranean Conference on Pattern Recognition and Artificial Intelligence; Springer: Cham, Switzerland, 2021; pp. 315–328. [Google Scholar]
  35. Breviglieri, P.; Erdem, T.; Eken, S. Predicting smart grid stability with optimized deep models. SN Comput. Sci. 2021, 2, 73. [Google Scholar] [CrossRef]
  36. Komatsu, H.; Kimura, O. Customer segmentation based on smart meter data analytics: Behavioral similarities with manual categorization for building types. Energy Build. 2023, 283, 112831. [Google Scholar] [CrossRef]
  37. Chicco, G.; Napoli, R.; Piglione, F. Comparisons among clustering techniques for electricity customer classification. IEEE Trans. Power Syst. 2006, 21, 933–940. [Google Scholar] [CrossRef]
  38. Chicco, G. Overview and performance assessment of the clustering methods for electrical load pattern grouping. Energy 2012, 42, 68–80. [Google Scholar] [CrossRef]
  39. Faria, P.; Spinola, J.; Vale, Z. Aggregation and remuneration of electricity consumers and producers for the definition of demand-response programs. IEEE Trans. Ind. Inform. 2016, 12, 952–961. [Google Scholar] [CrossRef]
  40. Li, D.; Chiu, W.; Sun, H.; Poor, H. Multiobjective optimization for demand side management program in smart grid. IEEE Trans. Ind. Inform. 2017, 14, 1482–1490. [Google Scholar] [CrossRef]
  41. Al Khafaf, N.; Jalili, M.; Sokolowski, P. Application of Deep Learning Long Short-Term Memory in Energy Demand Forecasting; Springer: Cham, Switzerland, 2019; pp. 31–42. [Google Scholar]
  42. Li, R.; Li, F.; Smith, N. Multi-resolution load profile clustering for smart metering data. IEEE Trans. Power Syst. 2016, 31, 4473–4482. [Google Scholar] [CrossRef]
  43. Verdú, S.; Garcia, M.; Senabre, C.; Marin, A.; Franco, F.C. Filtering, and identification of electrical customer load patterns through the use of self-organizing maps. IEEE Trans. Power Syst. 2006, 21, 1672–1682. [Google Scholar] [CrossRef]
  44. Coke, G.; Tsao, M. Random effects mixture models for clustering electrical load series. J. Time Ser. Anal. 2010, 31, 451–464. [Google Scholar] [CrossRef]
  45. McLoughlin, F.; Duffy, A.; Conlon, M. Characterising domestic electricity consumption patterns by dwelling and occupant socio-economic variables: An Irish case study. Energy Build. 2012, 48, 240–248. [Google Scholar] [CrossRef]
  46. Peel, D.; McLachlan, G. Robust mixture modelling using the t distribution. Stat. Comput. 2000, 10, 339–348. [Google Scholar] [CrossRef]
  47. Liu, C.; Rubin, D.M. Estimation of the t distribution using EM and its extensions, ECM and ECME. Stat. Sin. 1995, 5, 19–39. [Google Scholar]
  48. Wei, X.; Yang, Z. The infinite Student’s t-factor mixture analyzer for robust clustering and classification. Pattern Recognit. 2012, 45, 4346–4357. [Google Scholar] [CrossRef]
  49. Allili, M.; Bouguila, N.; Ziou, D. Finite general Gaussian mixture modeling and application to image and video foreground segmentation. J. Electron. Imaging 2008, 17, 013005. [Google Scholar]
  50. Elguebaly, T.; Bouguila, N. Bayesian learning of finite generalized Gaussian mixture models on images. Signal Process. 2011, 91, 801–820. [Google Scholar] [CrossRef]
  51. Elguebaly, T.; Bouguila, N. A nonparametric Bayesian approach for enhanced pedestrian detection and foreground segmentation. In Proceedings of the CVPR 2011 WORKSHOPS, Colorado Springs, CO, USA, 20–25 June 2011; pp. 21–26. [Google Scholar]
  52. Miller, J.; Thomas, J. Detectors for discrete-time signals in non-Gaussian noise. IEEE Trans. Inf. Theory 1972, 18, 241–250. [Google Scholar] [CrossRef]
  53. Farvardin, N.; Modestino, J. Optimum quantizer performance for a class of non-Gaussian memoryless sources. IEEE Trans. Inf. Theory 1984, 30, 485–497. [Google Scholar] [CrossRef]
  54. Gao, Z.; Belzer, B.; Villasenor, J.A. comparison of the Z, E/sub 8/, and Leech lattices for quantization of low-shape-parameter generalized Gaussian sources. IEEE Signal Process. Lett. 1995, 2, 197–199. [Google Scholar]
  55. Elguebaly, T.; Bouguila, N. Finite asymmetric generalized Gaussian mixture models learning for infrared object detection. Comput. Vis. Image Underst. 2013, 117, 1659–1671. [Google Scholar] [CrossRef]
  56. Elguebaly, T.; Bouguila, N. Model-based approach for high-dimensional non-Gaussian visual data clustering and feature weighting. Digit. Signal Process. 2015, 40, 63–79. [Google Scholar] [CrossRef]
  57. Hyvärinen, A.; Hoyer, P. Emergence of phase-and shift-invariant features by decomposition of natural images into independent feature subspaces. Neural Comput. 2000, 12, 1705–1720. [Google Scholar] [CrossRef] [PubMed]
  58. Farag, A.; El-Baz, A.; Gimel’farb, G. Precise segmentation of multimodal images. IEEE Trans. Image Process. 2006, 15, 952–968. [Google Scholar] [CrossRef] [PubMed]
  59. Bedingfield, S.; Alahakoon, D.; Genegedera, H.; Chilamkurti, N. Multi-granular electricity consumer load profiling for smart homes using a scalable big data algorithm. Sustain. Cities Soc. 2018, 40, 611–624. [Google Scholar] [CrossRef]
  60. Wang, Y.; Chen, Q.; Kang, C.; Xia, Q.; Luo, M. Sparse and redundant representation-based smart meter data compression and pattern extraction. IEEE Trans. Power Syst. 2016, 32, 2142–2151. [Google Scholar] [CrossRef]
  61. Yang, J.; Honavar, V. Feature subset selection using a genetic algorithm. Feature Extr. Constr. Sel. 1998, 13, 44–49. [Google Scholar]
  62. Elguebaly, T.; Bouguila, N. Simultaneous high-dimensional clustering and feature selection using asymmetric Gaussian mixture models. Image Vis. Comput. 2015, 34, 27–41. [Google Scholar] [CrossRef]
  63. Al-Otaibi, R.; Jin, N.; Wilcox, T.; Flach, P. Feature construction and calibration for clustering daily load curves from smart-meter data. IEEE Trans. Ind. Inform. 2016, 12, 645–654. [Google Scholar] [CrossRef]
  64. Iglesias, F.; Zseby, T.; Zimek, A. Absolute Cluster Validity. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 2096–2112. [Google Scholar] [CrossRef] [PubMed]
  65. Melzi, F.; Same, A.; Zayani, M.; Oukhellou, L.A. dedicated mixture model for clustering smart meter data: Identification and analysis of electricity consumption behaviors. Energies 2017, 10, 1446. [Google Scholar] [CrossRef]
  66. Wallace, C.; Dowe, D.L. MML Clustering of Multi-State, Poisson, von Mises circular and Gaussian distributions. Stat. Comput. 2000, 10, 73–83. [Google Scholar] [CrossRef]
  67. Wallace, C.; Freeman, P. Estimation and inference by compact coding. J. R. Stat. Soc. Ser. 1987, 49, 240–252. [Google Scholar] [CrossRef]
  68. Agusta, Y.; Dowe, D. Unsupervised learning of gamma mixture models using minimum message length. In Proceedings of the 3rd IASTED Conference on Artificial Intelligence and Applications, Benalma’dena, Spain, 8–10 September 2003; Acta Press: Benalmadena, Spain, 2003; pp. 457–462. [Google Scholar]
  69. Elguebaly, T.; Bouguila, N. Background subtraction using finite mixtures of asymmetric gaussian distributions and shadow detection. Mach. Vis. Appl. 2014, 25, 1145–1162. [Google Scholar] [CrossRef]
  70. Azam, M.; Bouguila, N. Multivariate-bounded Gaussian mixture model with minimum message length criterion for model selection. Expert Syst. 2021, 38, e12688. [Google Scholar] [CrossRef]
  71. Azam, M.; Bouguila, N. Multivariate bounded support laplace mixture model. Soft Comput. 2020, 24, 13239–13268. [Google Scholar] [CrossRef]
  72. Wallace, C.; Boulton, D. An information measure for classification. Comput. J. 1968, 11, 185–194. [Google Scholar] [CrossRef]
  73. Figueiredo, M.; Jain, A. Unsupervised learning of finite mixture models. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 381–396. [Google Scholar] [CrossRef]
  74. Conway, J.; Sloane, N. Sphere Packings, Lattices and Groups; Springer Science & Business Media: New York, NY, USA, 2013. [Google Scholar]
  75. Bouguila, N.; Ziou, D. High-dimensional unsupervised selection and estimation of a finite generalized Dirichlet mixture model based on minimum message length. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1716–1731. [Google Scholar] [CrossRef] [PubMed]
  76. Bezdek, J.; Pal, N. Some new indexes of cluster validity. IEEE Trans. Syst. Man, Cybern. Part B 1998, 28, 301–315. [Google Scholar] [CrossRef] [PubMed]
  77. Rousseeuw, P.S. Silhouettes: A graphical aid to the interpretation and validation of cluster analysis. J. Comput. Appl. Math. 1987, 20, 53–65. [Google Scholar] [CrossRef]
  78. Caliński, T.; Harabasz, J. A dendrite method for cluster analysis. Commun. Stat.-Theory Methods 1974, 3, 1–27. [Google Scholar] [CrossRef]
  79. Davies, D.L.; Bouldin, D.W. A cluster separation measure. IEEE Trans. Pattern Anal. Mach. Intell. 1979, 2, 224–227. [Google Scholar] [CrossRef]
  80. Allili, M. Wavelet modeling using finite mixtures of generalized Gaussian distributions: Application to texture discrimination and retrieval. IEEE Trans. Image Process. 2011, 21, 1452–1464. [Google Scholar] [CrossRef] [PubMed]
  81. Baldi, P.; Brunak, S.; Chauvin, Y.; Andersen, C.; Nielsen, H. Assessing the accuracy of prediction algorithms for classification: An overview. Bioinformatics 2000, 16, 412–424. [Google Scholar] [CrossRef] [PubMed]
  82. Schwarz, G. Estimating the dimension of a model. Ann. Stat. 1978, 6, 461–464. [Google Scholar] [CrossRef]
Figure 1. The Gaussian distribution symmetry problem.
Figure 1. The Gaussian distribution symmetry problem.
Sensors 23 08296 g001
Figure 2. Consumption profiles used to generate the synthetic datasets. (a) First synthetic dataset. (b) Second synthetic dataset.
Figure 2. Consumption profiles used to generate the synthetic datasets. (a) First synthetic dataset. (b) Second synthetic dataset.
Sensors 23 08296 g002
Figure 3. Mixture model’s log-likelihood function demonstration during the clustering of the synthetic datasets. (a) First synthetic dataset. (b) Second synthetic dataset.
Figure 3. Mixture model’s log-likelihood function demonstration during the clustering of the synthetic datasets. (a) First synthetic dataset. (b) Second synthetic dataset.
Sensors 23 08296 g003
Figure 4. The mixture models’ performance information during the clustering of the first real-life smart meter data. (a) Selection of the optimal number of mixture components using MML and the proposed model. (b) The log-likelihood functions of the mixture models used in the comparison.
Figure 4. The mixture models’ performance information during the clustering of the first real-life smart meter data. (a) Selection of the optimal number of mixture components using MML and the proposed model. (b) The log-likelihood functions of the mixture models used in the comparison.
Sensors 23 08296 g004
Figure 5. Household energy consumption segmentation demonstration of the first real-life smart meter dataset. (a) The average demand of all the energy consumers starting from 14 July 2009 to 31 December 2010. (b) The average demand of the optimal energy consumption clusters from 14 July 2009 to 31 December 2010.
Figure 5. Household energy consumption segmentation demonstration of the first real-life smart meter dataset. (a) The average demand of all the energy consumers starting from 14 July 2009 to 31 December 2010. (b) The average demand of the optimal energy consumption clusters from 14 July 2009 to 31 December 2010.
Sensors 23 08296 g005
Figure 6. Number of energy consumers in each cluster.
Figure 6. Number of energy consumers in each cluster.
Sensors 23 08296 g006
Figure 7. The UK Power Networks smart meter data clusters information. (a) Percentage of energy consumers in each cluster. (b) The distribution of tariffs across the resulting clusters.
Figure 7. The UK Power Networks smart meter data clusters information. (a) Percentage of energy consumers in each cluster. (b) The distribution of tariffs across the resulting clusters.
Sensors 23 08296 g007
Table 1. FSBAGGMM special cases.
Table 1. FSBAGGMM special cases.
Special CaseRequired Change in FSBAGGMM Parameters
Feature selection model based on
the Asymmetric Generalized Gaussian Mixture (FSAGGMM) [56] H ( X i d | k ) = 1
Feature selection model based on
the Bounded Asymmetric Gaussian Mixture (FSBAGMM) λ k d = 2
Feature selection model based on
the Asymmetric Gaussian Mixture (FSAGMM) [62] H ( X i d | k ) = 1 ,   λ k d = 2
Feature selection model based on
the Bounded Generalized Gaussian Mixture (FSBGGMM) σ r k d = σ l k d
Feature selection model based on
the Generalized Gaussian Mixture (FSGGMM) σ r k d = σ l k d ,   H ( X i d | k ) = 1
Feature selection model based on
the Bounded Gaussian Mixture (FSBGMM) σ r k d = σ l k d ,   λ k d = 2
Feature selection model based on
the Gaussian Mixture (FSGMM) σ r k d = σ l k d ,   λ k d = 2 ,   H ( X i d | k ) = 1
Feature selection model based on
the Bounded Laplace Mixture (FSBLMM) σ r k d = σ l k d ,   λ k d = 1
Feature selection model based on
the Laplace Mixture (FSLMM) σ r k d = σ l k d , λ k d = 1 ,   H ( X i d | k ) = 1
Asymmetric Generalized Gaussian Mixture Model (AGGMM) [55] H ( X i d | k ) = 1 ,   ω d = 1
Bounded Asymmetric Gaussian Mixture Model (BAGMM) λ k d = 2 ,   ω d = 1
Asymmetric Gaussian Mixture Model (AGMM) [69] H ( X i d | k ) = 1 ,   λ k d = 2 ,   ω d = 1
Bounded Generalized Gaussian Mixture Model (BGGMM) [18] σ r k d = σ l k d ,   ω d = 1
Generalized Gaussian Mixture Model (GGMM) [49] σ r k d = σ l k d ,   H ( X i d | k ) = 1 ,   ω d = 1
Bounded Gaussian Mixture Model (BGMM) [70] σ r k d = σ l k d ,   λ k d = 2 ,   ω d = 1
Gaussian Mixture Model (GMM) σ r k d = σ l k d ,   λ k d = 2 ,   H ( X i d | k ) = 1 ,   ω d = 1
Bounded Laplace Mixture Model (BLMM) [71] σ r k d = σ l k d ,   λ k d = 1 ,   ω d = 1
Laplace Mixture Model (LMM) σ r k d = σ l k d ,   λ k d = 1 ,   H ( X i d | k ) = 1 ,   ω d = 1
Table 2. Count of observations generated for the first synthetic dataset.
Table 2. Count of observations generated for the first synthetic dataset.
Gaussian White Noise ParametersProfile 1Profile 2Profile 3Profile 4Profile 5
μ = 0.001; σ = 0.2378370379371382
μ = 0.01; σ = 0.2349364356356355
μ = 0.1; σ = 0.2352360361359348
μ = 0.05; σ = 0.3354358359356353
μ = 0.01; σ = 0.3365353357350355
Table 3. Mixture models’ clustering performance evaluation using the first synthetic dataset.
Table 3. Mixture models’ clustering performance evaluation using the first synthetic dataset.
Performance Index (%)FSBAGGMMFSAGGMMBAGGMMAGGMM
ACC95.56994.33885.45882.804
TPR/Recall88.93585.83663.58956.953
PPV/Precision89.45888.14974.83870.500
MCC86.29182.92158.17051.104
F1-Score88.92285.84463.64457.011
TNR97.23196.46190.90689.245
NPV97.26396.59192.12890.942
FPR2.7693.5399.09410.755
FNR11.06514.16436.41143.047
FDR10.54211.85125.16229.500
Table 4. Mixture models’ clustering performance evaluation using the first synthetic dataset.
Table 4. Mixture models’ clustering performance evaluation using the first synthetic dataset.
Performance IndexOptimal Performance IndicatorFSBAGGMMFSAGGMMBAGGMMAGGMM
GOFMinimum3870.6837261.08316,397.63317,765.500
CHMaximum2081.8682046.4441594.2151405.947
SMaximum0.1070.1000.023−0.016
DBMinimum2.5492.6232.6612.503
DIMaximum0.2240.2190.2090.209
Xie and Benie IndexMinimum1.8711.8812.4462.698
Fowlkes MallowsMaximum0.7990.7550.6500.648
Log LossMinimum0.6250.9019.74112.138
EOEMinimum0.7300.7581.0221.032
JaccardMaximum0.8890.8580.6360.570
ROC AUCMaximum0.9310.9120.7730.731
V MeasureMaximum0.7550.7400.6600.639
RandMaximum0.9190.8990.8200.795
Normalized Mutual InformationMaximum0.7550.7400.6600.639
Mutual InformationMaximum1.2131.1810.9690.887
HomogeneityMaximum0.7540.7340.6020.551
Adjusted RandMaximum0.7490.6910.5240.497
Adjusted Mutual InfoMaximum0.7550.7400.6600.639
Table 5. Clusters using first synthetic dataset.
Table 5. Clusters using first synthetic dataset.
MethodFSBAGGMM
BIC7
AIC7
DI4
MML5
EoE5
GT5
Table 6. Clusters using second synthetic dataset.
Table 6. Clusters using second synthetic dataset.
MethodBAGGMM + FW
BIC6
AIC6
DI6
MML8
EoE8
GT8
Table 7. Count of observations generated for the second synthetic dataset.
Table 7. Count of observations generated for the second synthetic dataset.
Gaussian White Noise ParametersProfile 1Profile 2Profile 3Profile 4Profile 5Profile 6Profile 7Profile 8
μ = 0.001; σ = 0.2445448450444449447442455
μ = 0.01; σ = 0.2442449448448448452445448
μ = 0.1; σ = 0.2442452455449447443447445
μ = 0.05; σ = 0.3445448444451453447442450
μ = 0.01; σ = 0.3460459458468457455466457
Table 8. Mixture models’ clustering performance evaluation using the second synthetic dataset.
Table 8. Mixture models’ clustering performance evaluation using the second synthetic dataset.
Performance Index (%)FSBAGGMMFSAGGMMBAGGMMAGGMM
ACC91.85688.74688.48187.769
TPR/Recall67.45954.96953.86251.021
PPV/Precision66.48255.75356.40254.291
MCC63.81350.40249.90846.726
F1-Score67.42254.98353.92251.078
TNR95.34793.57093.41893.012
NPV95.45693.92193.92693.528
FPR4.6536.4306.5826.988
FNR32.54145.03146.13848.979
FDR33.51844.24743.59845.709
Table 9. Mixture models’ clustering performance evaluation using the second synthetic dataset.
Table 9. Mixture models’ clustering performance evaluation using the second synthetic dataset.
Performance IndexOptimal Performance IndicatorFSBAGGMMFSAGGMMBAGGMMAGGMM
GOFMinimum22,539.82036,474.84250,310.22548,011.423
CHMaximum2100.9551766.7971713.4501674.616
SMaximum0.0540.001−0.052−0.062
DBMinimum3.5634.9756.7676.738
DIMaximum0.2100.2130.2080.194
Xie and BenieMinimum2.8833.6193.6833.784
Fowlkes MallowsMaximum0.5740.4860.5180.503
Log LossMinimum3.29310.28712.61813.228
EOEMinimum0.6200.6370.6850.675
JaccardMaximum0.6740.5500.5390.511
ROC AUCMaximum0.8140.7430.7370.720
V MeasureMaximum0.6440.5650.5930.586
RandMaximum0.8810.8360.8310.821
Normalized Mutual InformationMaximum0.6440.5650.5930.586
Mutual InfoMaximum1.3031.0881.1141.093
HomogeneityMaximum0.6270.5230.5360.526
Adjusted RandMaximum0.5020.3840.4070.385
Adjusted Mutual InfoMaximum0.6440.5650.5930.585
Table 10. Identified optimal number of clusters for the real-life smart meter dataset.
Table 10. Identified optimal number of clusters for the real-life smart meter dataset.
Model Selection MethodFSBAGGMM
BIC3
AIC3
DI2
MML3
EoE4
Table 11. Mixture models’ clustering performance using the real-life smart meter dataset.
Table 11. Mixture models’ clustering performance using the real-life smart meter dataset.
Performance IndexMetric’s Optimal ValueFSBAGGMMFSAGGMMBAGGMMAGGMM
SMaximum0.2500.2160.2280.176
CHMaximum7.3775.8246.6715.594
DBMinimum16.95123.83220.62624.577
DIMaximum0.2530.2380.2490.224
Xie and BenieMinimum60.82172.96962.15773.319
EOEMinimum1.4601.7641.6131.822
Table 12. Consumption profile statistics for the year 2010.
Table 12. Consumption profile statistics for the year 2010.
Consumption Profile ClusterAverage Consumption (kWh)Annual Consumption ResponsibilityClusters’ Proportion
16536.77018.650%64.600%
216,117.19045.980%1.700%
312,394.57035.360%33.700%
Table 13. Identified optimal number of clusters for the second real-life smart meter dataset.
Table 13. Identified optimal number of clusters for the second real-life smart meter dataset.
Model Selection MethodFSBAGGMM
BIC4
AIC4
DI4
MML4
EoE2
Table 14. Mixture models’ clustering performance using the second real-life smart meter dataset.
Table 14. Mixture models’ clustering performance using the second real-life smart meter dataset.
Performance IndexMetric’s Optimal ValueFSBAGGMMFSAGGMMBAGGMMAGGMM
SMaximum0.3190.2880.2650.189
CHMaximum1984.8431078.837545.442243.243
DBMinimum1.0501.0752.5833.108
DIMaximum0.0270.0230.0190.012
Xie and BenieMinimum0.5500.7190.9391.283
EOEMinimum0.3150.4340.4420.453
Table 15. The mean values of the first seven smart meter data features.
Table 15. The mean values of the first seven smart meter data features.
Consumption ProfileOvernight RAPBreakfast RAPDaytime RAPEvening RAPMean STDSeasonal ScoreWD-WE Diff. Score
10.6860.9371.0411.3440.8100.8830.458
20.6641.0500.9561.4111.1271.0251.557
30.6720.9591.0111.3810.9742.0620.553
40.8600.9810.9161.2491.1694.4450.591
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Al-Bazzaz, H.; Azam, M.; Amayri, M.; Bouguila, N. Unsupervised Mixture Models on the Edge for Smart Energy Consumption Segmentation with Feature Saliency. Sensors 2023, 23, 8296. https://0-doi-org.brum.beds.ac.uk/10.3390/s23198296

AMA Style

Al-Bazzaz H, Azam M, Amayri M, Bouguila N. Unsupervised Mixture Models on the Edge for Smart Energy Consumption Segmentation with Feature Saliency. Sensors. 2023; 23(19):8296. https://0-doi-org.brum.beds.ac.uk/10.3390/s23198296

Chicago/Turabian Style

Al-Bazzaz, Hussein, Muhammad Azam, Manar Amayri, and Nizar Bouguila. 2023. "Unsupervised Mixture Models on the Edge for Smart Energy Consumption Segmentation with Feature Saliency" Sensors 23, no. 19: 8296. https://0-doi-org.brum.beds.ac.uk/10.3390/s23198296

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop