Next Article in Journal
Series Representations for Uncertain Fractional IVPs in the Fuzzy Conformable Fractional Sense
Previous Article in Journal
Statistics-Based Outlier Detection and Correction Method for Amazon Customer Reviews
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Singular Spectrum Analysis for Background Initialization with Spatio-Temporal RGB Color Channel Data

1
Department of Electronic Engineering, National Kaohsiung University of Science and Technology, Kaohsiung 80778, Taiwan
2
Department of Electronic Engineering, Ming Chi University of Technology, New Taipei City 24301, Taiwan
3
Institute of Photonics Engineering, National Kaohsiung University of Science and Technology, Kaohsiung 80778, Taiwan
*
Authors to whom correspondence should be addressed.
Submission received: 9 November 2021 / Revised: 4 December 2021 / Accepted: 5 December 2021 / Published: 7 December 2021
(This article belongs to the Section Signal and Data Analysis)

Abstract

:
In video processing, background initialization aims to obtain a scene without foreground objects. Recently, the background initialization problem has attracted the attention of researchers because of its real-world applications, such as video segmentation, computational photography, video surveillance, etc. However, the background initialization problem is still challenging because of the complex variations in illumination, intermittent motion, camera jitter, shadow, etc. This paper proposes a novel and effective background initialization method using singular spectrum analysis. Firstly, we extract the video’s color frames and split them into RGB color channels. Next, RGB color channels of the video are saved as color channel spatio-temporal data. After decomposing the color channel spatio-temporal data by singular spectrum analysis, we obtain the stable and dynamic components using different eigentriple groups. Our study indicates that the stable component contains a background image and the dynamic component includes the foreground image. Finally, the color background image is reconstructed by merging RGB color channel images obtained by reshaping the stable component data. Experimental results on the public scene background initialization databases show that our proposed method achieves a good color background image compared with state-of-the-art methods.

1. Introduction

Scene background initialization is a basic low-level process in video-processing applications, such as video segmentation [1], video compression [2], computational photography [3], and video surveillance [4,5] (e.g., tracking, counting). The background initialization is also known as background estimation, background reconstruction, and background generation. The task of background initialization can be described as follows: given a video, we need to construct a model that describes the clear background image despite the continued presence of moving objects. The background image may be valid for the entire video or updated in time if the background configuration changes due to illumination change or the displacement of background objects.
Figure 1a shows frames from the HighwayII sequence of the scene background initialization (SBI) database [6]. There is an appearance of moving objects in each frame, particularly cars. These frames are the input data of the background initialization model as described in Figure 1b. Using the proposed background initialization model, we can eliminate the appearance of moving objects to obtain a clean background, which is also known as the closest-to-ground-truth background, as shown in Figure 1c.
During the past two decades, many methods [1,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25] were proposed for the background initialization task. In general, these techniques can be classified into four main categories: pixel-based methods [1,7,8,9,10,11], iterative-based methods [12,13,14,15], low-rank/sparse data separation methods [16,17,18,19,20,21,22], and deep learning-based methods [23,24,25,26,27].
The first subcategory includes pixel-based methods where each pixel is processed individually over time. Chiu et al. [1] achieved the background by clustering the pixels. Pixels obtained from each location along its time axis are clustered according to their intensity variations. The pixel corresponding to the cluster that has a maximum probability more significant than a time-varying threshold is extracted as a background pixel. Maddalena and Petrosino [7] used a temporal median to compute the background pixel as the mean of the pixels at the same position across all the image sequences. The most well-known method is a mixture of Gaussians (MoG) proposed by Stauffer [9]. The background is modeled probabilistically at each pixel location by fitting MoG to the observed pixel valued in a recent temporal window. MoG decides whether each pixel is classified as background or foreground. More recently, Laugraud et al. [10] presented a method called LaBGen, which combined a pixel-wise temporal median filter and a patch-selection mechanism based on motion detection. In each frame, a background subtraction algorithm determines whether each pixel in the video belongs to the foreground or background. Tian et al. [11] introduced the block-level background modeling (BBM) algorithm to obtain video-coding background components. The BBM algorithm uses the residual gradient as the temporal information to distinguish the background blocks. BBM is used to consider the boundary difference, and the pixel smoothness process is handled using a weighted average of pixel temporal value.
The second subcategory includes iterative-based methods [12,13,14,15]. These methods usually consist of two stages. In the first stage, these methods detect static regions considered reference backgrounds. The background model is iteratively completed in the second stage based on suitable spatial consistency criteria. Hsiao and Leou [12] performed background initialization and foreground segmentation tasks based on motion estimation and computation of the correlation coefficient. Each block of the current frame is classified into four categories: background, still object, illumination change, and moving entity to exploit for the background updating phase. The static blocks, such as “background” and “illumination change”, are selected as the reference, and the remaining blocks are suitably used for the iterative completion of the background model. In [13], Torre and Black applied robust principal component analysis (RPCA) for separating the background and foreground to detect the outlier from video or image data. Firstly, the number of bases that preserve 55% of data energy is calculated using standard PCA. Then based on the obtained number of bases, RPCA is used for minimizing the vital energy function until convergence to receive the weight matrix. Finally, the weight matrix is used to detect outliers. Reitberger and Sauer [14] proposed a background-determining model based on an iterative singular value decomposition via singular vectors spanning a subspace of the image space. The method has a fast processing speed and can be applied in real-time applications. But it has difficulty handling challenges, such as intermittent motion. Recently, based on long-term background stable and short-term foreground changes of scenes, Chen et al. [15] adopted a Bayesian framework to classify the background and foreground.
The third subcategory includes low-rank/sparse data separation methods. The background information is considered low-rank information, and the remainder of the data represents both noises and moving objects. One of the first attempts to initialize the background in this subcategory was introduced by Candes et al. [16]. They perfectly separated a given video into a low-rank matrix and a sparse matrix by solving a very convenient convex program called principal component pursuit (PCP). However, PCP has several disadvantages for real-world videos, such as its time consumption and computational complexity. To overcome the limitations of the PCP method, many studies were proposed, such as Javed et al. [17] and Zhou et al. [18], which work well in specific environments. Ye et al. [19] presented a motion-assisted matrix restoration (MAMR) model for background-foreground separation of a video. In the MAMR model, the sparse matrix contains the foreground objects, and the low-rank matrix includes the background. A dense motion field is calculated and mapped into a weighting matrix for each frame, which indicates the likelihood that each pixel belongs to the background. In [20], Grosek and Kutz introduced the video dynamic mode decomposition (DMD) method for foreground and background separation. The DMD method decomposes video data into different dynamic modes, which are associated with Fourier frequencies. The frequencies near the origin do not change from frame to frame. Thus they are considered background components, and the terms with Fourier frequencies bounded away from the origin are foreground components. In [21], non-negative matrix factorization (NMF) was used to approximate a non-negative matrix A to a product of two non-negative, low-rank factor matrices W and H, where W contains background components and H contains foreground components. More recently, Kajo et al. [22] introduced a spatio-temporal, slice-based, singular value decomposition (SVD) method by organizing videos, such as tensors and seeks, to sparse them into different components. Each of these components, namely the moving object and the background, is represented by a few distinct significant eigenvalues. However, this proposal can be time-consuming to process over an ample space. Besides, it still has some limitations in the complex scenes, such as illumination variation, short video, and clutter.
The fourth subcategory includes deep learning-based methods [23,24,25,26,27]. These methods used the effectiveness of the deep learning model to automatically learn the background model. Ramirez-Quintana and Chacon-Murguia [23], based on self-organizing maps (SOMs) and cellular neural networks (CNNs), proposed a self-adaptive system named SOM-CNN. This system includes two neural network architectures called retinotopic SOM (RESOM) and neighbor threshold CNN (NTCNN) for video and motion analysis. The system can work with typical and complex scenarios in real time. Zhao et al. [24] proposed a background modeling method called the stacked multilayer self-organizing map background model (SMSOM-BM). This model can learn the background model of challenging scenarios and automatically determine most network parameters by considering every pixel and spatial consistency at each layer. Halfaoui et al. [25] proposed a CNN-based method to estimate the background component. This method is effective for challenges, such as dynamic backgrounds, illumination variation, and clusters. Yang et al. [26] proposed a deep neural network for background modeling. First, they used the temporal encoding to sample multiple frames from original sequential images with variable intervals, then they used a fully convolutional network to extract temporal and spatial information from frames. In the work by Gregorio et al. [27], the authors introduced a background initialization approach by weightless neural network. Each pixel is allied to an artificial weightless neural network that learns more frequently. This method is useful for processing long-term and live videos.
In the real world, background initialization still faces many challenges, such as lighting changes, the foreground occupying most of the frames, the automatic adjustment of the video camera, and objects moving heterogeneously (sometimes stationary, sometimes moving). To address these issues, we propose a novel method belonging to the low-rank/sparse data separation method named background initialization with singular spectrum analysis (BISSA). Firstly, the input image sequence is reorganized into a spatio-temporal data type useful for background–foreground separation tasks. Secondly, an adaptive background initialization algorithm for image sequences based on the SSA is proposed. Finally, to evaluate the effectiveness of our method, we compare our approach with some of the state-of-the-art techniques by doing experiments on a SBI [6] database. The experiment results show that our proposed method is more accurate and easier to apply in real-world applications.
The rest of the paper is organized as follows: Section 2 describes an overview of the SSA algorithm. Section 3 presents our proposed method. Finally, experimental results and discussion are summarized in Section 4, while conclusions and future work are represented in Section 5.

2. Singular Spectrum Analysis

In recent years, singular spectrum analysis (SSA) [28,29,30] has emerged as a powerful non-parametric tool to apply for analyzing and predicting time series data. This method aims to decompose the input data into a sum of different meaningful components, where these components can be grouped and merged based on their common properties to compose subsequent components. These grouped components indicate different groups of features of the original time series data. Currently, many researchers apply SSA in different areas, such as biomedical diagnostic tests [31], climatology [32], economics [33,34], signal processing [35], etc. A flowchart of SSA, consisting of the substages of decomposition and reconstruction, is shown in Figure 2.
As can be seen in Figure 2, the basic SSA algorithm consists of two isolated stages: decomposition and reconstruction stages. In the first stage, embedding and singular value decomposition steps are applied for the decomposition. In the last stage, eigentriple grouping and diagonal averaging steps are used to reconstruct the time series. For example, given non-zero time series X = ( f 1 , f 2 , , f K ) of length K , W is denoted as the window length and 1 < W < K ; L = K W + 1 . The SSA algorithm is described below:
Stage 1: Decomposition
Step 1: Embedding
Embedding is a standard procedure in time series analysis. Embedding can be regarded as a mapping that transfers a one-dimensional time series into a multidimensional series. By selecting a large window size, more information about the basis pattern of the time series is captured. Constructing the trajectory matrix F of the original time series X , which is a matrix of size W × L , gives:
F = ( f 1 f 2 f 3 f L f 2 f 3 f 4 f L + 1 f 3 f 4 f 5 f L + 2 f W f W + 1 f W + 2 f K ) W × L ,
where rows and columns of F are subseries of the original time series.
Step 2: Singular value decomposition (SVD)
This step computes the SVD of the trajectory matrix F sized W × L . By using SVD, matrix F can be decomposed into the product of three matrices: an orthogonal matrix U of size W × r, a diagonal matrix Σ of size r × r , and the transpose of another orthogonal matrix V of size r × L , where r is the rank of matrix F . In general, the SVD of trajectory matrix F can be written as:
F = U Σ V T = i = 1 r u i σ i v i T ,
where U = [ u 1 , u 2 , u 3 , , u r ] and V = [ v 1 , v 2 , v 3 , , v r ] are the column-orthonormal matrices, respectively (i.e., U T U = I and V T V = I ), and Σ = d i a g ( σ 1 , σ 2 , σ 3 , , σ r ) is a diagonal matrix containing the singular values (SVs) of F , where SVs are arranged in the descending order ( σ 1 σ 2 σ 3 σ r > 0 ) . The matrices F i = u i σ i v i T are called elementary matrices: they have rank-1. The collection ( u i , σ i , v i ) is called i -th eigentriple of matrix F .
Stage 2: Reconstruction
Step 3: Eigentriple grouping
This step can be used to analyze and determine the physical behavior of each component in the time series data. The purpose of the eigentriple grouping is to gather data based on their common properties. The different matrices of rank-1 acquired from applying the SVD of trajectory matrix F can be selected and gathered together. In that way, correctly clustered groups reflect other original time series data criteria. The grouping procedure separates the set of r eigentriples into m ( m r ) distinct subsets, and they are expressed as F G j = { F G 1 , F G 2 , , F G m } , where each F G j contains several F i and presents as:
F G j = [ f 11 , j f 21 , j f 31 , j f L 1 , j f 12 , j f 22 , j f 32 , j f L 2 , j f 13 , j f 23 , j f 33 , j f L 3 , j f 1 W , j f 2 ( W + 1 ) , j f 3 ( W + 1 ) , j f L K , j ] .
The progress of selecting the sets F G 1 , F G 2 , F G 3 , , F G m is called eigentriple grouping.
Step 4: Diagonal averaging
The final step is to perform the diagonal averaging on the matrices F G j where j = 1 , 2 , 3 , , m . This step converts grouped matrices F G j into a one-dimensional original time series via the diagonal averaging method. In particular, where F G j is a trajectory matrix grouped in step 3, the element f ˜ k j ,   k = 1 , 2 , , K of time series data S j is computed as the average of all elements on the minor diagonal kth of matrix F G j , which can be expressed as:
f ˜ k j = 1 k x + y = k + 1 1 x ,   y k f x y , j ,   k = 1 , 2 , , K .
The result of the reconstructed trajectory matrix along the diagonal averaging process is time series data of length K represented by:
S j = { f ˜ 1 j , f ˜ 2 j , f ˜ 3 j , , f ˜ K j } .

3. Background Initialization Using Singular Spectrum Analysis

Generally, background–foreground separation can be regarded as a matrix separation problem [16,36,37,38,39,40]. We can separate a video into two group components, one component that contains stable information and the remaining component that holds dynamic information. Constructing these components can be based on an eigentriple or a group of eigentriples. The background data (almost stable and highly correlated between frames) is contained in the static component, and the dynamic component usually represents the foreground data (moving objects or noise). The matrix separation problem can unify in a more general framework formulated as follows [16,36,37,38,39,40]:
X = S + ε D ,
where X is the input video data information, matrix S indicates the stable component, and ε D represents the dynamic component, respectively. These components are achieved by reconstructing one or a group of eigentriples of the trajectory matrix F . As a result, the stable and dynamic components are calculated as:
S = i = 1 τ u i σ i v i T ,
ε D = j = τ + 1 r u j σ j v j T ,
where 1 τ r and r is the rank of F . In this study, video X is stored in three matrices as spatio-temporal data M ( C ) . Trajectory matrices F ( C ) are constructed based on these spatio-temporal data matrices, where C { R ,   G ,   B } represents R, G, and B color channels. More details on how to construct video X as spatio-temporal data used as the input data for our background initialization system are introduced in the following subsections.

3.1. Storing a Video as Spatio-Temporal Data

A fundamental problem in mathematics is how to arrange data, through which they reveal the most critical information. By organizing the correct given data, we can solve our problem. In this section, we introduce a way of rearranging input video data to solve the problem of separating the background and moving objects. Spatio-temporal data [40] is a data type that contains both space and time characteristics of the original data. Spatial refers to space and temporal relates to time. Spatio-temporal data analysis is discovering patterns and knowledge from spatio-temporal data. A video can be considered a dynamic system with evolving frames, where each frame presents the system’s state. In this study, by flattening the color frames of a video as columns of matrices, we obtain spatio-temporal data.
A grayscale video is three-dimensional (3D) input data, which is the frame height (m), width (n), and time (k) with k frames of the video, as shown in Figure 3a. By reshaping each frame into a column of size 1 × a of a matrix of size k × a (where a = m × n ) , as shown in Figure 3b, we obtain the spatio-temporal data matrix. In this matrix, the correlation between pixels located at the same neighboring position between adjacent frames is preserved over time. Additionally, the video is mapped from 3D space into two-dimensional (2D) space, thereby reducing the complex computing.
Without loss of generality, X is assumed to be an original color video consisting of k frames with a resolution of a. To display multichannel images in the RGB space, 24 bits with 8 bits for each color channel is used. Firstly, each color frame is separated into three color channels, namely R, G, and B. Secondly, we flatten each color channel frame to one vector and arrange the vector side by side to form a spatio-temporal data matrix, called the color channel spatio-temporal matrix. Finally, we obtain three spatio-temporal data matrices corresponding to the three color channels. Based on the color frame, we construct three spatio-temporal data matrices. The process to flatten video’s frames to the color channel spatio-temporal data matrices is summarized in Algorithm 1, as follows:
Algorithm 1. Construct the Three-Color Channel Spatio-Temporal Data Matrices of the Video
Input:X is a color video consisting of k color frames, where each frame has a resolution of a = m × n.
Output: Three color channel spatio-temporal data matrices, M ( C ) ,   C   { R ,   G ,   B } , of the video.
1.f1, f2, … , fkExtract k frames fi of the video X.
2. f i ( C ) Separate frame fi into RGB color channel images, i = 1, 2, 3,.., k.
3. m i ( C ) flatten each f i ( C ) image into a vector column of size 1 × a.
4. M ( C ) Arrange the vector column m i ( C ) side by side to form color channel spatio-temporal data matrices.

3.2. Singular Spectrum Analysis for Background Initialization

This section presents the central part of our background initialization method using SSA in detail. We introduce how to apply SSA for the background initialization task, given that X is a color video sequence of k frames, where each frame has a resolution of a = m × n . Firstly, by using Algorithm 1, as discussed in Section 3.1, we receive three color channel spatio-temporal data matrices M ( C ) of size a × k ,   C { R ,   G ,   B } representing the R, G, or B color channel used, which can be written as:
M ( C ) = ( f 11 ( C ) f 12 ( C ) f 13 ( C ) f 1 k ( C ) f 21 ( C ) f 22 ( C ) f 23 ( C ) f 2 k ( C ) f 31 ( C ) f 32 ( C ) f 33 ( C ) f 3 k ( C ) f a 1 ( C ) f a 2 ( C ) f a 3 ( C ) f a k ( C ) ) a × k .
Embedding: We construct the trajectory matrices F ( C ) based on color channel spatio-temporal data matrices M ( C ) by embedding operator T . The dimensions of the matrices F ( C ) are determined by two window lengths,   W a and W k , where 1 W a a , 1 W k k , and 1 < W a W k < a k , then L a = ( a W a + 1 ) and L k = ( k W k + 1 ) . The input 2D matrix M ( C ) is organized into the matrix F ( C ) of size ( W a W k × L a L k ) as follows:
T ( M ( C ) ) = F ( C ) = ( T 1 ( C ) T 2 ( C ) T 3 ( C ) T L k ( C ) T 2 ( C ) T 3 ( C ) T 4 ( C ) T L k + 1 ( C ) T 3 ( C ) T 4 ( C ) T 5 ( C ) T L k + 2 ( C ) T W k ( C ) T W k + 1 ( C ) T W k + 2 ( C ) T k ( C ) ) W a ( C ) W k ( C ) × L a ( C ) L k ( C ) ,
where each T i ( C ) is a trajectory matrix of size W a ( C ) × L a ( C ) composed from the color channel spatio-temporal data matrix M ( C ) , such as:
T i ( C ) = ( f 1 i ( C ) f 2 i ( C ) f 3 i ( C ) f L a i ( C ) f 2 i ( C ) f 3 i ( C ) f 4 i ( C ) f L a + 1 i ( C ) f 3 i ( C ) f 4 i ( C ) f 5 i ( C ) f L a + 2 i ( C ) f W a i ( C ) f W a + 1 i ( C ) f W a + 2 i ( C ) f a i ( C ) ) W a ( C ) × L a ( C ) .
Decomposition: We perform SVD on the trajectory matrices F ( C ) to obtain sets of the rank-1 matrices.
F ( C ) = U ( C ) Σ ( C ) V ( C ) T = i = 1 r ( C ) u i ( C ) σ i ( C ) ( v i ( C ) ) T ,
where U ( C ) = [ u 1 ( C ) ,   u 2 ( C ) , , u r ( C ) ( C ) ] and V ( C ) = [ v 1 ( C ) ,   v 2 ( C ) , , v r ( C ) ( C ) ] are orthogonal matrices containing singular vectors, Σ ( C ) = d i a g ( σ 1 ( C ) ,   σ 2 ( C ) , , σ r ( C ) ( C ) ) contains sorted SVs in a non-increasing order, and r ( C ) is the rank of F ( C ) .
Grouping: The rank-one matrices are merged following general criteria; the aggregate of the rank-one matrices acquire the grouped matrices in N   ( N r ) groups.
F ( C ) = F G 1 ( C ) + F G 2 ( C ) + + F G N ( C ) ,
where F G m ( C ) = m = 1 N u m ( C ) σ m ( C ) ( v m ( C ) ) T .
Return to the object decomposition: The grouped matrices are transformed to the form of the input object by performing T 1 based on the diagonal averaging method, as described in Equation (4):
F ˜ G m ( C ) = T 1 ( F G m ( C ) ) ,  
where m = 1 , 2 , 3 , , N .

3.3. Grouping of Eigentriples

This section analyzes and determines the specific meaning of an eigentriple or a group of eigentriples in video data. The first step is to set a window length. The algorithm proposed in this study separates the set of eigentriples into two groups, as described in Equation (6). Both groups reconstruct output data, resulting in two reconstructed output component data for given input data, so we set all window lengths to 2 in this study.
We selected video sequences, namely Board, CaVignal, and IBMtest2, for analysis and observation. This experiment considers window length sizes 2, 4, and 10, respectively. Figure 4 presents the eigenvectors plot from trajectory matrices of the three videos, as we analyzed the data with different window length sizes. As shown in Figure 4, from top to bottom, the blue line represents the first component, and other color lines indicate the remaining components. We can see that the first eigenvector is always a constant over time. The eigenvalue represents the magnitude of the data, and the eigenvector indicates the direction of the data. Therefore, the first eigenvector represents the unchanged data component over time. Those are referred to as stable components ( S ), representing the background in the video. Because of that reason, we reconstructed the background in this first eigentriple-based video and dynamic component ( ε D ) obtained by remaining eigentriples.
In summary, we split the set of indices { 1 , 2 , , r } into two groups, namely a stable component and a dynamic component. The result of the step is the representation:
S = u 1 σ 1 v 1 T ,  
ε D = i = 2 r u i σ i v i T ,  
where r is the rank of trajectory matrix.

3.4. Proposed Method

From the arguments presented above, by using the first eigentriple of color channel spatio-temporal data matrices, we can construct the most effective background image of the given video. The remaining eigentriples are used to construct the foreground. The implementation of the main part of BISSA method can be summarized in Algorithm 2 as follows:
Algorithm 2. Initialize Background Using SSA
Input: Color channel spatio-temporal data matrices M ( C ) of video X.
Output: Obtain background and foreground images corresponding to each input color frame.
1.Construct trajectory matrix F ( C ) :
F ( C ) E q u a t i o n ( 10 ) ( M ( C ) )
2.Decompose the trajectory matrix F ( C ) into a sum of one-rank elementary matrices:
F ( C ) = U ( C ) Σ ( C ) V ( C ) T = i = 1 r ( C ) u i ( C ) σ i ( C ) ( v i ( C ) ) T ,
where   r ( C ) = r a n k ( F ( C ) ) .
3.Construct background component S ( C ) based on the first eigentriple group ( i = 1 ) :
S ( C ) = u 1 ( C ) σ 1 ( C ) ( v 1 ( C ) ) T
4.Construct the foreground component ε D ( C ) based on remaining eigentriple groups:
ε D ( C ) = i = 2 r ( C ) u i ( C ) σ i ( C ) ( v i ( C ) ) T  
5.Perform the diagonal averaging S ( C ) :
S ˜ ( C ) E q u a t i o n   14 ( S ( C ) )  
6.Perform the diagonal averaging ε D ( C ) :
ε ˜ D ( C ) E q u a t i o n   14 ( ε D ( C ) )  
7.Reshape the first column of S ˜ ( C ) to matrices sized m × n:
S ˜ ( C ) = r e s h a p e ( S ˜ ( C ) ,   m , n )  
8.Reshape the columns of ε ˜ D ( C ) to matrices sized m × n:
ε ˜ D , i ( C ) = r e s h a p e ( ε ˜ D , i ( C ) ,   m , n ) ,
where i = { 1 , 2 , 3 , , k } .
9.Merge three colors channels of S ˜ ( C ) to receive the background image of the video:
S b g = m e r g e ( S ˜ ( C ) )
10.Merge three colors channels of ε ˜ D ( C ) to receive the foreground image ε b g ,   i corresponding to i t h frame:
ε b g ,   i = m e r g e ( ε ˜ D , i ( C ) )
Given X is a video consisting of k color frames, where each frame has a resolution of a = m × n , after applying Algorithm 1, the three color channel spatio-temporal data matrices corresponding to three color channels are obtained. Each color channel spatio-temporal data matrix contains k columns and a rows. Each column corresponds to one frame, and each row contains k pixel values of the same pixel position in the video. By applying Algorithm 2 on the three color channel spatio-temporal data matrices separately, we process and find the relevance of all frames over time. In our proposed method, we use the eigentriples of the color channel spatio-temporal data matrix to construct two groups of matrices. The first group is constructed by using only the first eigentriple, which contains the background information of the video. The second group is built by using the remaining eigentriples, which include the foreground information. To receive the desired background images, we reshape each column with the first reconstruction matrix to a matrix of size m × n to obtain the chosen background images. The color background image is obtained by merging the three color channels. Moreover, k -achieved color background images are the same; thereby, only the first column of the first matrix is used to reconstruct the background image to save processing time. Experimental results that support our arguments are discussed in more detail in the next section.

4. Experimental Results and Discussion

In this section, to indicate the effectiveness of our proposed method, experiments for the background initialization problem are conducted on the most popular benchmark database named the SBI [6] database. We also compare our background initialization performance with some state-of-art background initialization, such as median [7], RPCA [13], dynamic mode decomposition (DMD) [20], non-negative matrix factorization (NMF) [21], and background estimation by WiSARD (Wilkes, Stonham and Aleksander Recognition Device) [27]. Finally, to assess the accuracy of the obtained background images against the ground truth images, we use several measurement metrics, such as structural similarity index (SSIM) [41], feature similarity index for image quality assessment (FSIM) [42], peak-signal-to-noise ratio (PSNR) [43], average gray-level error (AGE), and percentage of error pixels (pEPs) [44].

4.1. SBI Database

This database was introduced by L. Maddalena at the workshop on scene background modeling and initialization in 2016. The SBI database [6] consists of 14 different image sequences, namely Board, Candela_m1.10, CAVIAR1, CAVIAR2, CaVignal, Foliage, Hall&Monitor, HighwayI, HighwayII, HumanBody2, IBMtest2, People&Foliage, Snellen, and Toscana, as shown in Figure 5a. These sequences are composed of 6 to 740 frames, and their dimensions vary from 144 × 144 to 800 × 600 . Each image sequence is accompanied by a ground truth background image, as shown in Figure 5b. SBI was designed to evaluate existing and future background initialization algorithms. The image sequences in the SBI database are intended for different challenges in background initialization tasks, such as camera jitter and shadows challenge, intermittent motion challenge, clutter challenge, very short challenge, etc.
In the SBI database, the first one is the clutter challenge, where the objects appear almost to cover the entire background, such as the Board, People&Foliage, Foliage, and Snellen sequences. The HighwayI and HighwayII sequences have many cars that are constantly moving. These sequences include challenges, such as shadows and camera jitter. The Candela_m1.10 sequence presents a scenario where a man appears with his bag and leaves the scene with nothing. The CaVignal sequence is challenging because the man appears and retains position in more than 60% of the frames before leaving. Some sequences include the challenges of intermittent motion, such as Candela_m1.10, CAVIAR1, CAVIAR2, CaVignal, Hall&Monitor, and People&Foliage. The other sequences, HumanBody2 and IBMtest2, contain basic challenges. Finally, the Toscana sequence consists of only 6 frames, which presents a short video challenge. Our goal is to propose a background estimation method to obtain the factual background of a given video.

4.2. Evaluation and Result

Our paper proposes an efficient method for the background initialization task. As discussed in Section 3, given X is a color video sequence of k frames f i ,   i = 1 ,   2 ,   ,   k with a resolution of a = m × n , the background initialization algorithm based on SSA is proposed. Firstly, color frames are split into three color channels, f i ( C ) , C   { R ,   G ,   B } representing the R, G, or B color channels, then flattened to a vector column of color channel spatio-temporal data matrices M ( C ) . These matrices contain both space and time characteristic information of the original video. In M ( C ) matrices, the correlation between pixels located at the same position between adjacent frames is preserved over time. Next, M ( C ) matrices are decomposed by SSA to find the eigentriples for constructing the stable and dynamic components. The stable component containing the background information is computed by grouping the first eigentriple, and the remaining eigentriples construct the dynamic component. Finally, by reshaping an arbitrary column of S ˜ ( C ) to matrices of size m × n , then merging the three color channels, we receive a corresponded background image of the video. Similarly, by reshaping the columns of ε ˜ D ( C ) to matrices of size m × n , then merging the three color channels, we obtain a sequence of foreground images corresponding to each video frame.
Figure 6 displays achieved background and foreground images corresponding to the frames in HighwayI, IBMtest2, CAVIAR2, and HighwayII sequences by using our proposed method. The background images presented in the second row and the foreground images corresponding to the frames are illustrated in the third row. These videos contain several challenges, such as intermittent motion, shadows, camera jitter, and basic. Using our proposed method, for each video containing k frames, we can obtain k background image and k foreground image. This study focused on the background initialization task; however, we also obtained impressive foreground results, as shown in the third row of Figure 6. As can be seen, all moving objects are eliminated from the original frame’s image. However, all the background images are the same, as shown in the second row of Figure 6. Therefore, we only need to construct a video’s background image by reshaping one column of the stable component to an image to reduce processing time.
Our proposed algorithm can obtain background and foreground images of a video. However, we focus on handling the background initialization in this paper. To show the greater effectiveness of BISSA, we compare the proposed method with several existing methods, such as median [7], RPCA [13], DMD [20], NMF [21], and BEWiS [27].
Figure 7 shows the ground truth background images and obtained background images using different methods of 14 different video sequences in the SBI dataset. The Toscana sequence in the SBI database includes only six frames that represent the challenge of very short videos. Therefore the convergence criterion of the Toscana sequence is not met in RPCA, which means we cannot compute the matrix that contains the weighting of each pixel in the training data. Figure 7a presents ground truth background images of 14 video sequences in the SBI dataset. Figure 7b–g illustrate the background images obtained by using BEWiS, median, RPCA, DMD, NMF, and our proposed method, respectively. As seen, our proposed method obtains a clear background image in most cases, such as the Board, Candela_m1.10, CAVIAR1, CAVIAR2, Hall&Monitor, HighwayI, HighwayII, HumanBody2, IBMtest2, and Snellen image sequences. For the CaVignal image sequence, the obtained background image is not as expected because the man appears and retains position in more than 60% of the frames before leaving, much like the People&Foliage video sequence, in which the result is not expected because the people and trees appear in 338 out of 341 frames. For the Toscana video sequence, the results are not as good as expected due to too few frames (only six frames) and the object appears to occupy the majority of the video. In summary, our proposal achieves positive results on the basic challenge, intermittent motion challenge, camera jitter challenge, and shadows challenge, but struggles a little in handling clutter video and a very short video sequence. However, the obtained results are still really good when compared to other methods, such as RPCA, DMD, and NMF.
To assess the accuracy of the obtained background images against the ground truth images, we use five measurement metrics: SSIM [41], FSIM [42], PSNR [43], AGE, and pEPs [44], to measure the similarity between the two images. These measurement metrics are image-to-image metrics measuring the visual correctness of an estimated background image against a ground truth background image. These methods exploit different aspects of image quality evaluation, thus leading to an extensive comprehensive evaluation of the obtained result. Table 1 summarizes the rank of values and preference of these measurement metrics. As can be seen in Table 1, for the SSIM, FSIM, and PSNR measurement metrics, the higher obtained values demonstrate a higher similarity between the two images. On the contrary, for the AGE and pEPs measurement metrics, the lower of the obtained values show a higher similarity between the obtained backgrounds and ground truth images. A summary is presented in Table 1.
A summary is presented in Table 2 that highlights the best values of the corresponding metrics in bold. As shown in Table 2, the BISSA method gets high performance in most videos when we use the SSIM, FSIM, and PSNR metrics to evaluate. With the pEPs metric, our proposed method gets high performance in most videos except the Board, CaVignal, and Snellen sequences. BEWiS is the best performer in the Foliage, but BISSA is still better than the RPCA, DMD, median, and NMF methods. With the AGE metric, our proposed method gets high performance in most videos except Canlenda_m1.10, CAVIAR1, and HumanBody2. However, the BISSA is still better than the remaining methods. When using the PSNR metric to evaluate results, our proposed method also gets high performance in most videos except Canlenda_m1.10 and Snellen. In most videos, the backgrounds obtained by our proposed method are very similar to the ground truths.

5. Conclusions

This study proposes an effective background initialization algorithm for image sequences. By storing color frame sequences of the video into color channel spatio-temporal data matrices, we can preserve the correlation between pixels located at the same position between adjacent frames over time. Next, the SSA method was applied to these spatio-temporal data. Then, the stable component is constructed by using the first eigentriple, which is the component that holds the color background image. In addition, encouraging results of the foreground component were obtained based on the remaining eigentriples. The experiment results on the most popular public scene background initialization database demonstrate our proposed method’s effectiveness. The obtained background image is compared to the ground truth background image by the five most common metrics: SSIM, FSIM, PSNR, AGE, and pEPs. The results proved that our study achieved some positive results, especially in dealing with basic challenges, cluster challenges, intermittent motion challenges, camera jitter challenges, and intense shadow challenges. In addition, the results also show that our proposed method achieves a good color background image when compared with state-of-the-art techniques, such as BEWiS, median, RPCA, DMD, and NMF. However, videos recorded with few frames (less than 20 frames) and intermittent object motion challenges (such as CaVignal sequence) remain open challenges. Moreover, computing the background from the first eigentriple only obtains a good estimation of the background, but is not optimal to get a reasonable estimate of the foreground. In the future, we will continue to improve our method to achieve better background and accurately detect moving objects in video.

Author Contributions

H.D.L. developed methodology, software coding, and wrote the original draft. J.-W.W. guided the research direction and edited the paper. T.N.L. designed the experiments and edited the paper. Y.-S.L. contributed to editing the paper. All authors discussed the results and contributed to the final manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Ministry of Science and Technology, Taiwan, Grant MOST 170-2218-E-992-310.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors appreciate the support from National Kaohsiung University of Science and Technology in Taiwan.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chiu, C.C.; Ku, M.Y.; Liang, L.W. A robust object segmentation system using a probability-based background extraction algorithm. IEEE Trans. Circuits Syst. Video Technol. 2010, 20, 518–528. [Google Scholar] [CrossRef]
  2. Paul, M. Efficient video coding optimal compression plane and background modeling. IET Image Process. 2012, 6, 1311–1318. [Google Scholar] [CrossRef]
  3. Granados, M.; Seidel, H.P.; Lensch, H.P. Background estimation from non-time sequence images. In Proceedings of the Graphics Interface 2008, Toronto, ON, Canada, 28–30 May 2008; pp. 33–40. [Google Scholar]
  4. Bouwmans, T. Tranditional and recent approaches in background modeling for foreground detection: An overview. Comput. Sci. Rev. 2014, 11, 31–66. [Google Scholar] [CrossRef]
  5. Maddalena, L.; Petrosino, A.; Russo, F. People counting by learning their appearance in a multi-view camera environment. Pattern Recognit. Lett. 2014, 36, 125–134. [Google Scholar] [CrossRef]
  6. Maddalena, L.; Petrosino, A. Towards benchmarking scene background initialization. In Proceedings of the International Conference on Image Analysis and Processing, Genoa, Italy, 8 September 2015; pp. 469–476. [Google Scholar]
  7. Maddalena, L.; Petrosino, A. The 3dSOBS+ algorithm for moving object detection. Comput. Vis. Image Underst. 2014, 122, 65–73. [Google Scholar] [CrossRef]
  8. Xu, Z.; Min, B.; Cheung, R.C.C. A robust background initialization algorithm with superpixel motion detection. Signal Process. Image Commun. 2019, 71, 1–12. [Google Scholar] [CrossRef] [Green Version]
  9. Stauffer, C.; Grimson, W.E.L. Adaptive background mixture models for real-time tracking. In Proceedings of the 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Fort Collins, CO, USA, 23–25 June 1999; pp. 246–252. [Google Scholar]
  10. Laugraud, B.; Piérard, S.; Van Droogenbroeck, M. LaBgen: A method based on motion detection for generating the background of a scene. Pattern Recognit. Lett. 2017, 96, 12–21. [Google Scholar] [CrossRef] [Green Version]
  11. Tian, L.; Wang, H.; Zhou, Y.; Peng, C. Video big data in smart city: Background construction and optimization for surveillance video processing. Future Gener. Comput. Syst. 2018, 86, 1371–1382. [Google Scholar] [CrossRef]
  12. Hsiao, H.H.; Leou, J.J. Background initialization and foreground segmentation for bootstrapping video sequences. EURASIP J. Image Video Process. 2013, 2013, 12. [Google Scholar] [CrossRef] [Green Version]
  13. De la Torre, F.; Black, M.J. Robust principal component analysis for computer vision. In Proceedings of the Eighth IEEE International Conference on Computer Vision, Vancouver, BC, Canada, 7–14 July 2001. [Google Scholar]
  14. Reitberger, G.; Sauer, T. Background subtraction using Adaptive Singular Value Decomposition. J. Math. Imaging Vis. 2020, 62, 1159–1172. [Google Scholar] [CrossRef]
  15. Chen, Z.; Wang, R.; Zhang, Z.; Wang, H.; Xu, L. Background foreground interaction for moving object detection in dynamic scenes. Inf. Sci. 2019, 483, 65–81. [Google Scholar] [CrossRef]
  16. Candès, E.J.; Li, X.; Ma, Y.; Wright, J. Robust principal component analysis? J. ACM 2011, 58, 1–37. [Google Scholar] [CrossRef]
  17. Javed, S.; Mahmood, A.; Bouwmans, T.; Jung, S.K. Spatiotemporal low-rank modeling for complex scene background initialization. IEEE Trans. Circuits Syst. Video Technol. 2016, 28, 1315–1329. [Google Scholar] [CrossRef]
  18. Zhou, X.; Yang, C.; Yu, W. Moving object detection by detecting contiguous outliers in the low-rank representation. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 597–610. [Google Scholar] [CrossRef] [Green Version]
  19. Ye, X.; Yang, J.; Sun, X.; Li, K.; Hou, C.; Wang, Y. Foreground background separation from video clips via motion-assisted matrix restoration. IEEE Trans. Circuits Syst. Video Technol. 2015, 25, 1721–1734. [Google Scholar] [CrossRef]
  20. Grosek, J.; Kutz, J.N. Dynamic mode decomposition for real-time background/foreground separation in video. arXiv 2014, arXiv:1404.7592. [Google Scholar]
  21. Xu, Y.; Yin, W.; Wen, Z.; Zhang, Y. An alternating direction algorithm for matrix completion with nonnegative factors. Font. Math. China 2012, 7, 365–384. [Google Scholar] [CrossRef]
  22. Kajo, I.; Kamel, N.; Ruicheck, Y.; Malik, A.S. SVD-based tensor-completion technique for background initialization. IEEE Trans. Image Process. 2018, 27, 3114–3126. [Google Scholar] [CrossRef] [PubMed]
  23. Ramirez-Quintana, J.A.; Chacon-Murguia, M.I. Self-adaptive SOM-CNN neural system for dynamic object detection in normal and complex scenarios. Pattern Recognit. 2015, 48, 1137–1149. [Google Scholar] [CrossRef]
  24. Zhao, Z.; Zhang, X.; Fang, Y. Stacked multilayer self-organizing map for background modeling. IEEE Trans. Image Process. 2015, 24, 2841–2850. [Google Scholar] [CrossRef]
  25. Halfaoui, I.; Bouzaraa, F.; Urfalioglu, O. CNN-Based Initial Background Estimation. In Proceedings of the 23rd International Conference on Pattern Recognition (ICPR), Cancun, NM, USA, 4–8 December 2016. [Google Scholar]
  26. Yang, L.; Li, J.; Luo, Y.; Zhao, Y.; Cheng, H.; Li, J. Deep background modeling using fully convolutional network. IEEE Trans. Intell. Transp. Syst. 2018, 19, 254–262. [Google Scholar] [CrossRef]
  27. De Gregorio, M.; Giordano, M. Background estimation by weightless neural networks. Parttern Recognit. Lett. 2017, 96, 55–65. [Google Scholar] [CrossRef]
  28. Broomhead, D.S.; King, G.P. Extracting qualitative dynamics from experemental data. Phys. D Nonlinear Phenom. 1986, 20, 217–236. [Google Scholar] [CrossRef]
  29. Elsner, J.B.; Tsonis, A.A. Singular Spectrum Analysis: A New Tool in Time Series Analysis; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  30. Golyandina, N.; Zhigljavsky, A. Singular Spectrum Analysis for Time Series, 2nd ed.; Springer Briefs in Statistics: Berlin, Germany, 2020. [Google Scholar]
  31. Ghodsi, M.; Hassani, H.; Sanei, S.; Hicks, Y. The use of noise information for detection of temporomandibular disorder. Biomed. Signal Process. Control 2009, 4, 79–85. [Google Scholar] [CrossRef]
  32. Yurova, A.; Bobylev, L.P.; Zhu, Y.; Davy, R.; Korzhikov, A.Y. Atmospheric heat advection in the Kara Sea region under main synoptic processes. Int. J. Climatol. 2019, 39, 361–374. [Google Scholar] [CrossRef] [Green Version]
  33. Arteche, J.; García-Enríquez, J. Singular sepectrum analysis for signal extraction in strochastic volatility models. Econom. Stat. 2017, 1, 85–98. [Google Scholar]
  34. Lahmiri, S. Minute-ahead stock price forecasting based on singular spectrum analysis and support vector regression. Appl. Math. Comput. 2018, 320, 444–451. [Google Scholar] [CrossRef]
  35. Golyandina, N. Particularities and commonalities of singular spectrum analysis as a method of time series analysis and signal processing. WIREs Comput. Stat. 2020, 12, e1487. [Google Scholar] [CrossRef]
  36. Zhou, Z.; Li, X.; Wright, J.; Candes, E.; Ma, Y. Stable principal component pursuit. In Proceedings of the 2010 IEEE International Symposium on Information Theory, Austin, TX, USA, 12–18 June 2010. [Google Scholar]
  37. Oreifej, O.; Li, X.; Shah, M. Simultaneous video stabilization and moving object detection in turbulence. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 450–462. [Google Scholar] [CrossRef] [Green Version]
  38. Zhang, L.; Chen, Z.; Zheng, M.; He, X. Robust non-negative matrix factorization. Front. Electr. Electron. Eng. China 2011, 6, 192–200. [Google Scholar] [CrossRef]
  39. Gonzalez, C.G.; Absil, O.; Absil, P.A.; Van Droogenbroeck, M.; Mawet, D.; Surdej, J. Low-rank plus sparse decomposition for exoplanet detection in direct-imaging ADI sequences-The LLSG algorithm. Astron. Astrophys. 2016, 589, A54. [Google Scholar] [CrossRef] [Green Version]
  40. Han, J.; Pei, J.; Kamber, M. Data Mining: Concepts and Techniques, 3rd ed.; Morgan Kaufmann: San Mateo, CA, USA, 2012. [Google Scholar]
  41. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  42. Zhang, L.; Zhang, L.; Mou, Z.; Zhang, D. FSIM: A feature similarity index for image quality assessment. IEEE Trans. Image Process. 2011, 20, 2378–2386. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Gonzalez, R.C.; Wood, R.E. Digital Image Processing, 3rd ed.; Prentice-Hall: Upper Saddle River, NJ, USA, 2007. [Google Scholar]
  44. Jodoin, P.M.; Maddalena, L.; Petrosino, A.; Wang, Y. Extensive benchmark and survey of modeling methods for scene background initialization. IEEE Trans. Image Process. 2017, 26, 5244–5256. [Google Scholar] [CrossRef]
Figure 1. The background initialization task: (a) image sequence, (b) background initialization model, and (c) desired background.
Figure 1. The background initialization task: (a) image sequence, (b) background initialization model, and (c) desired background.
Entropy 23 01644 g001
Figure 2. Flowchart of singular spectrum analysis algorithm.
Figure 2. Flowchart of singular spectrum analysis algorithm.
Entropy 23 01644 g002
Figure 3. The process of storing a video as a spatio-temporal data matrix: (a) a color video sequence of k frames with a resolution of m × n and (b) a spatio-temporal data matrix where each column represents one frame of the original video.
Figure 3. The process of storing a video as a spatio-temporal data matrix: (a) a color video sequence of k frames with a resolution of m × n and (b) a spatio-temporal data matrix where each column represents one frame of the original video.
Entropy 23 01644 g003
Figure 4. Eigenvectors plot of trajectory matrices for SBI database: (a) Board sequence, (b) CaVignal sequence, and (c) IBMtest2 sequence. The blue line indicates the direction of the first eigentriple group, and other color lines indicate the direction of the remaining eigentriple groups.
Figure 4. Eigenvectors plot of trajectory matrices for SBI database: (a) Board sequence, (b) CaVignal sequence, and (c) IBMtest2 sequence. The blue line indicates the direction of the first eigentriple group, and other color lines indicate the direction of the remaining eigentriple groups.
Entropy 23 01644 g004
Figure 5. Fourteen different image sequences in the SBI dataset, namely Board, Candela_m1.10, CAVIAR1, CAVIAR2, CaVignal, Foliage, Hall&Monitor, HighwayI, HighwayII, HumanBody2, IBMtest2, People&Foliage, Snellen, and Toscana (from left to right): (a) the frames of image sequences and (b) the ground truth background images of image sequences, respectively.
Figure 5. Fourteen different image sequences in the SBI dataset, namely Board, Candela_m1.10, CAVIAR1, CAVIAR2, CaVignal, Foliage, Hall&Monitor, HighwayI, HighwayII, HumanBody2, IBMtest2, People&Foliage, Snellen, and Toscana (from left to right): (a) the frames of image sequences and (b) the ground truth background images of image sequences, respectively.
Entropy 23 01644 g005
Figure 6. The background (second row) and foreground (third row) images corresponding to the original frames (first row) obtained by using our proposed method in different video sequences in the SBI dataset: (a) HighwayI sequence, (b) IBMtest2 sequence, (c) CAVIAR2 sequence, and (d) HighwayII sequence.
Figure 6. The background (second row) and foreground (third row) images corresponding to the original frames (first row) obtained by using our proposed method in different video sequences in the SBI dataset: (a) HighwayI sequence, (b) IBMtest2 sequence, (c) CAVIAR2 sequence, and (d) HighwayII sequence.
Entropy 23 01644 g006
Figure 7. The ground truth background images and obtained background images by using different methods of 14 different image sequences in the SBI dataset, namely Board, Candela_m1.10, CAVIAR1, CAVIAR2, CaVignal, Foliage, Hall&Monitor, HighwayI, HighwayII, HumanBody2, IBMtest2, People&Foliage, Snellen, and Toscana (from left to right): (a) ground truth background images, (b) BEWiS, (c) median, (d) RPCA, (e) DMD, (f) NMF, and (g) BISSA.
Figure 7. The ground truth background images and obtained background images by using different methods of 14 different image sequences in the SBI dataset, namely Board, Candela_m1.10, CAVIAR1, CAVIAR2, CaVignal, Foliage, Hall&Monitor, HighwayI, HighwayII, HumanBody2, IBMtest2, People&Foliage, Snellen, and Toscana (from left to right): (a) ground truth background images, (b) BEWiS, (c) median, (d) RPCA, (e) DMD, (f) NMF, and (g) BISSA.
Entropy 23 01644 g007aEntropy 23 01644 g007b
Table 1. Evaluation metrics.
Table 1. Evaluation metrics.
Eval. Met.NameRange of ValuePreference
SSIM [41]Structural similarity index[0–1]higher
FSIM [42]Feature similarity index for image quality assessment[0–1]higher
PSNR [43]Peak-signal-to-noise ratio[0–infinity]higher
AGE [44]Average gray-level error[0–255]lower
pEPs [44]Percentage of error pixels[0–1]lower
Table 2. Average results of the various methods in the SBI database.
Table 2. Average results of the various methods in the SBI database.
Method SSIM FSIM AGE pEPs PSNR
BoardBISSA0.730.8330.000.5316.5
RPCA0.630.7740.460.6413.4
NMF0.660.8334.880.6015.2
DMD0.590.7439.290.3812.3
BEWiS0.720.8218.660.2816.5
Median0.710.8021.740.2516.4
Candela_m1.10BISSA0.950.936.120.0325.76
RPCA0.940.944.920.0327.2
NMF0.940.954.640.0427.6
DMD0.940.944.630.0424.8
BEWiS0.950.963.670.0328.66
Median0.920.935.320.0425.78
CAVIAR1BISSA0.960.976.360.0727.8
RPCA0.910.9412.640.1221.3
NMF0.920.9616.700.2722.1
DMD--131.850.93-
BEWiS0.970.973.850.4527.27
Median0.900.939.180.0817.8
CAVIAR2BISSA0.950.993.200.0132.9
RPCA0.960.9812.870.0824.8
NMF0.960.9813.840.1124.5
DMD0.970.972.490.0130.1
BEWiS0.980.990.780.0447.61
Median0.960.963.400.0225.08
CaVignalBISSA0.880.8812.140.1020.04
RPCA0.740.7926.780.3615.3
NMF0.850.9014.890.1419.1
DMD0.770.8414.50.1416.5
BEWiS0.960.9812.760.1020.01
Median0.830.8712.90.1016.90
FoliageBISSA0.570.7236.460.7015.63
RPCA0.570.7441.810.5614.1
NMF0.690.8135.960.6015.4
DMD0.340.6350.950.6411.6
BEWiS0.870.9111.80.1715.38
Median0.600.7232.300.5423.75
Hall_monitorBISSA0.940.936.700.0328.3
RPCA0.910.948.060.0625.1
NMF0.930.954.680.0328.1
DMD0.890.936.030.0423.2
BEWiS0.920.933.621.4327.17
Median0.900.932.70.9926.46
HighwayIBISSA0.950.957.70.0229.08
RPCA0.830.9046.680.9414.3
NMF0.850.9242.830.9815.1
DMD0.660.7618.390.2918.9
BEWiS0.940.952.100.4654.49
Median0.890.931.420.1540.14
HighwayIIBISSA0.940.974.50.00333.32
RPCA0.930.964.300.0130.5
NMF0.940.973.570.00533.3
DMD0.810.899.760.1122.4
BEWiS0.940.962.190.4134.62
Median0.910.911.720.3134.66
HumanBody2BISSA0.950.969.710.1224.22
RPCA0.920.949.510.0822.5
NMF0.950.968.050.0825.9
DMD0.850.8913.00.1318.8
BEWiS0.950.984.261.5027.97
Median0.950.974.550.0131.96
People&FoliageBISSA0.740.848.580.6814.02
RPCA0.620.777.930.0612.2
NMF0.670.827.220.0713.0
DMD0.460.6910.10.0710.2
BEWiS0.660.7834.570.4012.45
Median0.660.7831.360.3813.60
IBMtest2BISSA0.950.9741.650.1125.5
RPCA0.910.9441.760.4727.02
NMF0.900.9442.480.6126.8
DMD0.880.9253.960.5721.4
BEWiS0.940.9643.981.5025.65
Median0.860.9148.910.1522.14
SnellenBISSA0.770.87530.8313
RPCA0.700.8250.740.8212.9
NMF0.760.8639.920.7515.0
DMD0.570.7661.40.7310.9
BEWiS0.760.8054.630.5225.75
Median0.690.7262.200.6213.65
ToscanaBISSA0.870.9316.180.2721.23
RPCA-----
NMF0.300.7770.030.9410.3
DMD0.760.8622.580.2816.82
BEWiS0.850.9210.370.1220.87
Median0.860.948.710.1320.67
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Le, H.D.; Le, T.N.; Wang, J.-W.; Liang, Y.-S. Singular Spectrum Analysis for Background Initialization with Spatio-Temporal RGB Color Channel Data. Entropy 2021, 23, 1644. https://0-doi-org.brum.beds.ac.uk/10.3390/e23121644

AMA Style

Le HD, Le TN, Wang J-W, Liang Y-S. Singular Spectrum Analysis for Background Initialization with Spatio-Temporal RGB Color Channel Data. Entropy. 2021; 23(12):1644. https://0-doi-org.brum.beds.ac.uk/10.3390/e23121644

Chicago/Turabian Style

Le, Huy D., Tuyen Ngoc Le, Jing-Wein Wang, and Yu-Shan Liang. 2021. "Singular Spectrum Analysis for Background Initialization with Spatio-Temporal RGB Color Channel Data" Entropy 23, no. 12: 1644. https://0-doi-org.brum.beds.ac.uk/10.3390/e23121644

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop