Next Article in Journal
Relativistic Corrections to the Higgs Boson Decay into a Pair of Vector Quarkonia
Next Article in Special Issue
A New Detection Function Model for Distance Sampling Based on the Burr XII Model
Previous Article in Journal
The Role of the Hidden Color Channel in Some Interesting Dibaryon Candidates
Previous Article in Special Issue
Analysis of Adaptive Progressive Type-II Hybrid Censored Dagum Data with Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Procedure for Change-Point Estimation Using Quantile Regression Model with Asymmetric Laplace Distribution

School of Mathematics and Statistics, Shandong University, Weihai 264209, China
Submission received: 3 January 2023 / Revised: 28 January 2023 / Accepted: 2 February 2023 / Published: 8 February 2023
(This article belongs to the Special Issue Symmetry in Statistics and Data Science)

Abstract

:
The usual mean change-point detecting method based on normal linear regression is not robust to heavy-tailed data with potential outlying points. We propose a robust change-point estimation procedure based on a quantile regression model with asymmetric Laplace error distribution and develop a non-iterative sampling algorithm from a Bayesian perspective. The algorithm can generate independently and identically distributed samples approximately from the posterior distribution of the position of the change-point, which can be used for statistical inferences straightforwardly. The procedure combines the robustness of quantile regression and the computational efficiency of the non-iterative sampling algorithm. A simulation study is conducted to illustrate the performance of the procedure with satisfying findings, and finally, real data is analyzed to show the usefulness of the algorithm by comparison with the usual change-point detection method based on normal regression.

1. Introduction

Regression models incorporating change-points have been popular in many application areas, such as finance (Chen [1]), medicine research (Kim et al. [2]), biology (Chen and Wang [3], Acitas and Senoglu [4]), climate (Reeves et al. [5]). In these documents, change-point detection in the normal linear regression model is probably the most popular. Chen and Gupta [6] first used the Schwarz Information Criterion (SIC) model selection method to detect the mean and variance change-points in the normal linear regression model. Chen and Wang [3] applied the normal mean change-point model to detect DNA copy number variations using the SIC method. Chen et al. [7] compared four methods for estimating single or multiple change-points in a normal regression model when both the error variance and regression coefficients change simultaneously at the unknown point. Meanwhile, many robust change-point detecting models with heavy-tailed distributions have been developed to deal with data with outliers. Osorio and Galea [8] discussed the regression coefficient change-point model with the Student-t error assumption and designed the SIC method to detect the position of the change-point. Moreover, in the Student t regression framework, Lin et al. [9] proposed a kind of Bayesian methodology to detect the variance change-point and analyze the U.S. stock market data. Kang et al. [10] extended Lin’ work to linear models with variance change-point by the assumption of a class of symmetric heavy-tailed errors which includes the Student-t distribution. Lu and Chang [11] developed a robust algorithm to detect regression coefficient change-point by incorporating the Student-t distribution with the expectation-maximization (EM) algorithm and the fuzzy classification procedure. Yang [12] designed a robust mean change-point detection procedure by employing the Laplace regression model via the EM algorithm. Alin et al. [13] proposed a robust F-test for change-point detection in linear regression models with a non-normal error structure.
The quantile regression (QR) model, pioneered by Koenker and Bassett [14], has been powerful in comprehensively assessing the effects of explanatory covariates at different quantiles of response outcomes. In contrast to the usual mean regression model, the QR model is less sensitive to outliers in the data; thus, when the data comes from some heavy-tailed distribution, the inferences based on QR are more credible than the ones derived from the usual mean regression. Since a set of quantiles often provides a more detailed description of the response distribution than the usual mean, QR offers a practically important alternative to the classical mean regression. Koenker [15] provides a detailed introduction to QR in theories, applications and computation. In recent years, many authors have devoted themselves to studying the change-point problem in the QR framework; Qu [16] proposed several tests for structural change in regression quantiles, such as fluctuation type statistic and Wald type statistic. Zhang et al. [17] developed a procedure for testing change-point due to covariate threshold in regression quantiles. Gabriela [18] considered the estimation problem in a nonlinear quantile model with change-points. Guo and He [19] discussed the quantile auto-regression change-point issue from a Bayesian perspective and developed the monte carlo markov chain (MCMC) algorithm. Although MCMC algorithms are widely used in Bayesian inference for complex data, we should mention that it is an iterative sampling algorithm; thus, the samples generated from the full conditional distributions are not independent, and the convergence assessment is first needed before inferences using these samples. Despite several well-known convergence criteria, determining when the markov chain achieves convergence is still a challenging task in Bayesian analysis.
Tan et al. [20] discovered an equation called the Inverse Bayes formula (IBF) and developed it into a sampling algorithm named IBF sampler, which is highly useful in doing bayesian sampling. Compared to the classical MCMC sampling, this sampler possesses two main advantages. First, it is non-iterative and thus fast. Second, it can draw independently and identically distributed (i.i.d) samples exactly or approximately from the target posterior distribution, and the samples can be used for statistical inferences immediately without convergence diagnosis. The main idea of the IBF sampler is to use the EM algorithm to obtain an optimal importance sampling density and then use the sampling/importance resampling algorithm to get i.i.d. samples from the target distribution. To get a comprehensive understanding of the IBF sampler, one can refer to Tan et al. [21], which gave a detailed introduction to IBF, including its extensions and wide applications. In recent years, the applications of the IBF algorithm appear prosperous, to name a few, Yang and Yuan [22,23] applied the idea of IBF to deal with quantile regression model and scale mixture normal regression model, respectively. Tsionas [24] used the IBF sampler to research financial areas with stochastic volatility models. As to change-point detection, Tian et al. [25] was the first to apply the non-iterative Bayesian method to deal with change-point problem, and the authors identified the position of change-points in a sequence of hemolytic uremic syndrome using IBF sampler. Recently, Yang and Yuan [26,27] used the idea of an IBF sampler to detect the position of change-point in normal mean and normal regression coefficient, respectively.
Motivated by Tan et al. [20,21], in this article, a non-iterative Bayesian sampling algorithm to detect the position of regression coefficient change-point based on the QR model is discussed. The procedure involves several key steps. Firstly, the representation of asymmetric Laplace distribution (ALD) makes it possible to transfer the observed likelihood into a missing data structure by augmenting a latent exponential distributed variable, which is crucial for implementing the EM algorithm and IBF sampler. Then, the best posterior importance sampling distribution to overlap the target density is obtained by using the posterior mode from the EM algorithm. Finally, the sampling IBF and sampling/importance resampling algorithm are combined to get i.i.d. samples approximately from the observed posterior distributions. Simulation studies are conducted to evaluate the robustness of the procedure, and real data are analyzed to illustrate the usefulness of the algorithm by comparison with the classical method to detect the position of the change-point in regression models. The novelty of this paper lies in that it is the first time to apply the IBF algorithm to focus on a change-point problem based on a quantile regression model, which combines the robustness of quantile regression in detecting change-point and the computational effectiveness of the IBF sampler.

2. Quantile Regression (QR) Model

Let Y denote a response variable and X the related explanatory vector with p components, and suppose the observations are ( y i , x i ) , i = 1 , , n . We introduce the linear quantile regression model as
y i = x i β q + ε i , ( i = 1 , , n ) ,
where β q is a unknown parameters vector of interest, ε i is the error term whose distribution is restricted to have the q-th quantile equal to zero, and x i denotes the transpose of x i . So, β q quantified the q-th conditional quantile regression coefficient of y i given x i , for i = 1 , , n , and hereafter, β q is simplified as β . The estimation β ^ of β is the one minimizing
S ( β ) = i = 1 n ρ q ( y i x i β ) ,
with ρ q ( · ) denoting the loss function ρ q ( u ) = u { q I ( u < 0 ) } , and here I ( · ) is the indicator function. Note that when q = 0.5 , the loss function is the symmetric absolute loss, which leads to the Laplace regression or median regression. Realizing that ρ q ( u ) is not differentiable at zero, β ^ cannot be obtained in a closed form. In documents, some elaborate methods, such as the interior point algorithm and the Majorization–Minimization algorithm are developed to obtain β ^ , see Koenker and Park [28], Hunter and Lange [29]. Many authors have realized that minimizing S ( β ) is equivalent to maximizing the likelihood function of a linear regression model with errors following the asymmetric Laplace distribution (ALD ( 0 , σ , q ) ), whose density function is given by
f ( u ) = q ( 1 q ) σ exp { ρ q ( u σ ) } , u R .
In the literature, Koenker and Machado [30] studied the likelihood ratio test statistics quantile regression coefficients with ALD assumption, Yu and Moyeed [31] proposed the statistical inference of QR from a Bayesian perspective based on the ALD error distribution. Using a novel location-scale mixture representation of ALD, Reed and Yu [32], Kozumi and Kobayashi [33] developed an efficient and easily implemented Gibbs sampling method for quantile regression in the Bayesian framework. This representation has been widely used in the latest parameter quantile regression studies, such as longitudinal data and variable selection. To name a few, Zhou et al. [34] and Tian et al. [35] proposed the EM algorithm for computing the MLE or posterior mode of regression coefficients in the linear quantile regression model, Luo et al. [36] studied the quantile regression model for longitudinal data using a Gibbs sampler. Alhamzawi and Yu [37], Ji et al. [38] studied the variable selection problems in quantile regression, tobit and binary quantile regression. Tian et al. [39] discussed the inferences problems of a mixture of linear quantile regression using the EM algorithm. We notice that recently, Gallardo et al. [40] introduced a new parametric quantile regression model for asymmetric response variables with a power skew–normal distribution, and Reyes et al. [41] proposed a kind of quantile regression model with a generalization of the Student-t distribution as response variables, although these two distributions are different from the asymmetric Laplace distribution in this paper, the ideas of re-parametrization of these two distributions are similar to the mixture representation of ALD.

3. Quantile Regression Model with Change-Point

In this paper, we suppose a change-point exists in the q-th quantile regression coefficients at an unknown position k, such that
y i = x i β 1 + ε i , i = 1 , 2 , , k , y i = x i β 2 + ε i , i = k + 1 , , n , ε i ALD ( 0 , 1 , q ) , i = 1 , 2 , , n .
where β 1 = ( β 1 , , β p ) and β 2 = ( β 1 * , , β p * ) are the q-th conditional quantile regression coefficients for the first k observations and the last n k observations, respectively. For simplicity, the scale parameter σ in the ALD error distribution is set to 1. Our principal interest involves the determination of the unknown change-point position k and other parameters from a Bayesian perspective. Denote θ = ( β 1 , β 2 , k ) and observation vector y = ( y 1 , , y n ) , then the likelihood function of θ based on y is
L ( θ | y ) = i = 1 k q ( 1 q ) exp { ρ q ( y i x i β 1 ) } × i = k + 1 n q ( 1 q ) exp { ρ q ( y i x i β 2 ) } exp [ i = 1 k ρ q ( y i x i β 1 ) + i = k + 1 n ρ q ( y i x i β 2 ) ] .
By combining this likelihood with some prior distributions, we can get the kernel of the posterior distribution, but the complexity of the ALD makes the posterior inference difficult. So Reed and Yu [32], Kozumi and Kobayashi [33], and Zhou et al. [34] represented the ALD ( 0 , σ , q ) into a normal-exponential mixing distribution as Proposition 1.
Proposition 1
(Mixture Representation of the Asymmetric Laplace Distribution). Let z be an exponential variable with mean σ and u a standard normal variable. If a random variable ϵ follows the asymmetric Laplace distribution with density (2), then we can represent ε as a location-scale mixture of normals given by ε = θ z + τ z u , where
θ = 1 2 q q ( 1 q ) , τ 2 = 2 q ( 1 q ) .
Based on this representation, the quantile regression model with a single change-point (3) becomes
y i | z i N ( x i β 1 + θ z i , τ 2 z i ) , i = 1 , , k . y i | z i N ( x i β 2 + θ z i , τ 2 z i ) , i = k + 1 , , n . z i Exp ( 1 ) , i = 1 , 2 , , n .
where z i , i = 1 , , n are mutually independent. By introducing latent vector z = ( z 1 , , z n ) , the likelihood function based on complete data ( y , z ) appears as
L ( β 1 , β 2 , k | y , z ) = i = 1 k 1 2 π τ 2 z i exp { ( y i x i β 1 θ z i ) 2 2 τ 2 z i z i } × i = k + 1 n 1 2 π τ 2 z i exp { ( y i x i β 2 θ z i ) 2 2 τ 2 z i z i } .

4. Bayesian Inferences and IBF Algorithm

4.1. The Prior and Conditional Distributions

To emphasize the influence of data on the position of change-point and other parameters, we adopt a set of independent non-informative priors as follows
π ( β 1 ) 1 , π ( β 2 ) 1 , π ( k ) 1 n 2 p + 1 , k = p , , n p ,
here the prior of k is a discrete uniform distribution on { p , p + 1 , , n p } . The implementation of the IBF algorithm involves several conditional distributions, which are π ( β 1 | y , z , k ) , π ( β 2 | y , z , k ) , π ( k | y , z ) , and π ( z | y , β 1 , β 2 , k ) . These distributions are all proportional to the joint conditional distribution of the parameters, that is
π ( β 1 , β 2 , k | y , z ) = π ( β 1 | y , z , k ) π ( β 2 | y , z , k ) π ( k | y , z ) L ( β 1 , β 2 , k | y , z ) π ( β 1 ) π ( β 2 ) π ( k ) exp { 1 2 i = 1 k ( y i x i β 1 θ z i ) 2 τ 2 z i } × exp { 1 2 i = k + 1 n ( y i x i β 2 θ z i ) 2 τ 2 z i } × exp { i = 1 n z i } I ( p k n p ) ,
with I ( · ) representing the indicator function. Denote
X 1 = ( x 1 , , x k ) , W 1 = diag ( τ 2 z 1 , , τ 2 z k ) ,
z 1 = ( z 1 , , z k ) , y 1 = ( y 1 , , y k )
and
b 1 = ( X 1 W 1 1 X 1 ) 1 X 1 W 1 1 ( y 1 θ z 1 ) ,
we have
π ( β 1 | y , z , k ) exp ( β 1 b 1 ) X 1 W 1 1 X 1 ( β 1 b 1 ) 2 ,
consequently,
β 1 | y , z , k N ( b 1 , ( X 1 W 1 1 X 1 ) 1 ) .
Likewise, let
X 2 = ( x k + 1 , , x n ) , W 2 = diag ( τ 2 z k + 1 , , τ 2 z n ) ,
z 2 = ( z k + 1 , , z n ) , y 2 = ( y k + 1 , , y n ) ,
and
b 2 = ( X 2 W 2 1 X 2 ) 1 X 2 W 2 1 ( y 2 θ z 2 ) ,
we have
π ( β 2 | y , z , k ) exp ( β 2 b 2 ) X 2 W 2 1 X 2 ( β 2 b 2 ) 2 ,
consequently,
β 2 | y , z , k N ( b 2 , ( X 2 W 2 1 X 2 ) 1 ) .
Next, by integrating β 1 , β 2 from the joint distribution π ( β 1 , β 2 , k | y , z ) in Equation (4), one can get the conditional distribution of the position of change-point k based on complete data ( y , z ) , that is for k = p , p + 1 , , n p ,
π ( k | y , z ) | X 1 W 1 1 X 1 | 1 / 2 | X 2 W 2 1 X 2 | 1 / 2 exp ( S 1 + S 2 + 2 i = 1 n z i 2 ) ,
with
S i = ( y i θ z i X i b i ) W i 1 ( y i θ z i X i b i ) , i = 1 , 2 .
Finally, we present the conditional distribution of latent vector z , that is
π ( z | y , β 1 , β 2 , k ) L ( β 1 , β 2 , k | y , z ) i = 1 k z i 1 2 exp { 1 2 [ ( y i x i β 1 ) 2 τ 2 1 z i + 2 τ 2 + θ 2 τ 2 z i ] } × i = k + 1 n z i 1 2 exp { 1 2 [ ( y i x i β 2 ) 2 τ 2 1 z i + 2 τ 2 + θ 2 τ 2 z i ] } ,
explicitly, z 1 , , z n are mutually conditional independent, and for i = 1 , 2 , , k ,
z i | y i , β 1 GIG ( 1 2 , ( y i x i β 1 ) 2 τ 2 , 2 τ 2 + θ 2 τ 2 ) ,
for i = k + 1 , , n ,
z i | y i , β 2 GIG ( 1 2 , ( y i x i β 2 ) 2 τ 2 , 2 τ 2 + θ 2 τ 2 ) ,
where GIG ( λ , a , b ) denotes the generalized inverse Gaussian distribution with density
f ( x | λ , a , b ) = ( b / a ) λ / 2 2 K λ ( a b ) x λ 1 exp { 1 2 ( a x 1 + b x ) } ,
and K λ ( · ) presents the modified Bessel function of the third kind in Barndorff-Nielsen and Shephard [42].

4.2. EM Algorithm to Compute the Posterior Modes

To implement the IBF sampler, we first use the EM algorithm to find the observed posterior mode θ 0 = ( β 1 0 , β 2 0 , k 0 ) of π ( β 1 , β 2 , k | y ) and get the optimal importance sampling density π ( z | y , θ 0 ) . The log-posterior based on complete data ( y , z ) is that
l ( β 1 , β 2 , k | y , z ) = 1 2 i = 1 k ( y i x i β 1 θ z i ) 2 τ 2 z i 1 2 i = k + 1 n ( y i x i β 2 θ z i ) 2 τ 2 z i i = 1 n z i .
Based on Zhou et al. [34] and Tian et al. [35], we get
γ j i = E ( z i | y i , β j ) = τ 2 θ 2 + 2 τ 2 + | y i x i β j | θ 2 + 2 τ 2 , δ j i = E ( 1 z i | y i , β j ) = θ 2 + 2 τ 2 | y i x i β j | , j = 1 , 2 .
Suppose the initial value of θ is ( β 1 ( 0 ) , β 2 ( 0 ) , k ( 0 ) ) , and the ( t 1 ) -th iteration values is θ ( t 1 ) = ( β 1 ( t 1 ) , β 2 ( t 1 ) , k ( t 1 ) ) , the EM algorithm in the t-th iteration involves two steps:
E step: Omitting the terms without relation to θ , the Q function in the E step is the conditional expectation of l ( θ | y , z ) with respect to the conditional distribution π ( z | y , θ ( t 1 ) ) , which is presented as
Q ( θ | y , θ ( t 1 ) ) = E [ l ( θ | y , z ) | y , θ ( t 1 ) ] = 1 τ 2 [ i = 1 k δ 1 i ( t 1 ) ( y i x i β 1 ) 2 2 θ i = 1 k ( y i x i β 1 ) + θ 2 i = 1 k γ 1 i ( t 1 ) ] 1 τ 2 [ i = k + 1 n δ 2 i ( t 1 ) ( y i x i β 2 ) 2 2 θ i = k + 1 n ( y i x i β 2 ) + θ 2 i = k + 1 n γ 2 i ( t 1 ) ] [ i = 1 k γ 1 i ( t 1 ) + i = k + 1 n γ 2 i ( t 1 ) ] ,
with
γ j i ( t 1 ) = τ 2 θ 2 + 2 τ 2 + | y i x i β j ( t 1 ) | θ 2 + 2 τ 2 , δ j i ( t 1 ) = θ 2 + 2 τ 2 | y i x i β j ( t 1 ) | , j = 1 , 2 .
M step: We maximize the Q function with respect to θ . For fixed k, let W 1 ( t 1 ) = diag ( δ 11 ( t 1 ) , , δ 1 k ( t 1 ) ) , and
W 2 ( t 1 ) = diag ( δ 2 , k + 1 ( t 1 ) , , δ 2 , n ( t 1 ) ) ,
then we get
β 1 ( t ) ( k ) = ( X 1 W 1 ( t 1 ) X 1 ) 1 ( X 1 W 1 ( t 1 ) y 1 θ 1 k ) ,
β 2 ( t ) ( k ) = ( X 2 W 2 ( t 1 ) X 2 ) 1 ( X 2 W 2 ( t 1 ) y 2 θ 1 n k ) ,
with 1 k indicating the k dimension vector with each element equal 1. By denoting the convergence values as β ^ 1 ( k ) and β ^ 2 ( k ) , we can obtain
k ^ = arg max p k n p l ( β ^ 1 ( k ) , β ^ 2 ( k ) , k | y ) = arg min p k n p { S 1 ( k ) + S 2 ( k ) } ,
with
S 1 ( k ) = i = 1 k ρ q ( y i x i β ^ 1 ( k ) ) , S 2 ( k ) = i = k + 1 n ρ q ( y i x i β ^ 2 ( k ) ) .
and then, we use samples ( y i , x i ) , i = 1 , , k ^ and ( y i , x i ) , i = k ^ + 1 , , n to obtain the posterior mode β ^ 1 and β ^ 2 via another EM algorithm, respectively. After convergence, we get θ 0 = ( β 1 0 , β 2 0 , k 0 ) , the mode of posterior distribution π ( θ | y ) .

4.3. IBF Sampler

Based on the following expression
π ( θ | y ) = π ( θ | y , z ) π ( z | y ) d z = π ( β 1 | y , z , k ) π ( β 2 | y , z , k ) π ( k | y , z ) π ( z | y ) d z ,
in order to get samples from π ( θ | y ) , we first need to draw samples from π ( z | y ) . The IBF sampler proposed by Tan et al. [20,21] shows that
π ( z | y ) π ( z | y , θ 0 ) π ( θ 0 | y , z ) ,
where π ( z | y , θ 0 ) is an importance sampling density or proposal density in the sampling/importance resampling algorithm, with θ 0 denoting the posterior mode of π ( θ | y ) obtained by EM algorithm in Section 4.2. The IBF sampling algorithm follows by
Step 1. Based on (14), for i = 1 , , k 0 , draw i.i.d. samples
{ z i ( j ) } j = 1 J GIG ( 1 2 , ( y i x i β 1 0 ) 2 τ 2 , 2 τ 2 + θ 2 τ 2 ) ,
and based on (15), for i = k 0 + 1 , , n , draw i.i.d. samples
{ z i ( j ) } j = 1 J GIG ( 1 2 , ( y i x i β 2 0 ) 2 τ 2 , 2 τ 2 + θ 2 τ 2 ) .
Denote z ( j ) = ( z 1 ( j ) , z 2 ( j ) , , z n ( j ) ) , then { z ( j ) } j = 1 J are the samples from π ( z | y , θ 0 ) .
Step 2. Calculate the weights
λ j = 1 π ( θ 0 | y , z ( j ) ) l = 1 J 1 π ( θ 0 | y , z ( l ) ) , j = 1 , 2 , J ,
where
π ( θ 0 | y , z ( j ) ) = π ( β 1 0 | y , z ( j ) , k 0 ) π ( β 2 0 | y , z ( j ) , k 0 ) π ( k 0 | y , z ( j ) ) ,
is obtained from (8), (10) and (12) with z replaced by z ( j ) .
Step 3. Choose a subset from { z ( j ) } j = 1 J via resampling without replacement from the Step discrete distribution on { z ( j ) } j = 1 J with probabilities { λ j } j = 1 J to obtain an i.i.d. sample of size L ( < J ) approximately from π ( z | y ) , denoted by { z ( j l ) } l = 1 L .
Step 4. Denote
z 1 ( j l ) = ( z 1 ( j l ) , , z k ( j l ) ) , z 2 ( j l ) = ( z k + 1 ( j l ) , , z n ( j l ) ) ,
W 1 ( j l ) = diag ( τ 2 z 1 ( j l ) ) , W 2 ( j l ) = diag ( τ 2 z 2 ( j l ) ) ,
b 1 ( j l ) = ( X 1 W 1 ( j l ) 1 X 1 ) 1 X 1 W 1 ( j l ) 1 ( y 1 θ z 1 ( j l ) ) ,
b 2 ( j l ) = ( X 2 W 2 ( j l ) 1 X 2 ) 1 X 2 W 2 ( j l ) 1 ( y 2 θ z 2 ( j l ) ) ,
and
S 1 ( j l ) = ( y 1 θ z 1 ( j l ) X 1 b 1 ( j l ) ) W 1 ( j l ) 1 ( y 1 θ z 1 ( j l ) X 1 b 1 ( j l ) ) ,
S 2 ( j l ) = ( y 2 θ z 2 ( j l ) X 2 b 2 ( j l ) ) W 2 ( j l ) 1 ( y 2 θ z 2 ( j l ) X 2 b 2 ( j l ) ) .
For k = p , p + 1 , , n p , let
w k l = | X 1 W 1 j l 1 X 1 | 1 / 2 | X 2 W 2 j l 1 X 2 | 1 / 2 exp ( S 1 ( j l ) + S 2 ( j l ) + 2 i = 1 n z i j l 2 ) ,
and
p k l = w k l k = p n p w k l , k = p , , n p .
For l = 1 , , L , sample k ( l ) from { p , , n p } with probabilities { p k l , k = p , , n p } based on (12).
Step 5. Denote
X 1 l = ( x 1 , , x k ( l ) ) , X 2 l = ( x k ( l ) + 1 , , x n ) ,
y 1 l = ( y 1 , , y k ( l ) ) , y 2 l = ( y k ( l ) + 1 , , y n ) ,
z 1 ( l ) = ( z 1 ( j l ) , , z k ( l ) ( j l ) ) , z 2 ( l ) = ( z k ( l ) + 1 ( j l ) , , z n ( j l ) ) ,
W 1 ( l ) = diag ( z 1 ( l ) ) , W 2 ( l ) = diag ( z 2 ( l ) ) ,
b 1 ( l ) = ( X 1 l W 1 ( l ) 1 X 1 l ) 1 X 1 l W 1 ( l ) 1 ( y 1 l θ z 1 ( l ) ) ,
b 2 ( l ) = ( X 2 l W 2 ( l ) 1 X 2 l ) 1 X 2 l W 2 ( l ) 1 ( y 2 l θ z 2 ( l ) ) ,
for l = 1 , , L , based on (9), generate
β 1 ( l ) N b 1 ( l ) , ( X 1 l W 1 ( l ) 1 X 1 l ) 1 ,
and based on (11), generate
β 2 ( l ) N b 2 ( l ) , ( X 2 l W 2 ( l ) 1 X 2 l ) 1 .
Denote θ ( l ) = ( β 1 ( l ) , β 2 ( l ) , k ( l ) ) , then, { θ ( l ) } l = 1 L are the i.i.d. samples approximately from π ( θ | y ) , which can be used for statistical inferences immediately without convergence diagnosis.

5. Simulation Studies

In this section, we conduct simulations to investigate the performance of the proposed approach to detect the position of the change-point by comparing it with the normal mean change-point detecting method proposed by Chen [1]. We generate n = 200 observations with a change-point at position k from the following model:
y i = β 1 x 1 i + β 2 x 2 i + ε i , i = 1 , , k , y i = β 1 * x 1 i + β 2 * x 2 i + ε i , i = k + 1 , , n .
where β 1 = ( β 1 , β 2 ) = ( 1 , 1 ) , β 2 = ( β 1 * , β 2 * ) = ( 2 , 3 ) , and x j i U ( 2 , 2 ) for j = 1 , 2 and i = 1 , , n . Here k is set as 40, 60, 80, 100, 120, 140, and 160, respectively, and the following four error distributions are considered:
  • Normal distribution: ε i N ( 0 , 1 ) ,
  • Laplace distribution: ε i L ( 0 , 1 ) ,
  • t distribution with three degree of freedom: ε i t ( 3 ) ,
  • Cauchy distribution.
In these distributions, cases 2–4 are heavy-tailed and often used to describe the outlier situations, and especially case 4, the Cauchy distribution is extremely heavy-tailed. The quantile levels to estimate the position of change-point are set as 0.25, 0.5 and 0.75, and the experiment is replicated 500 times. In each replication, firstly, a position k ( 0 ) = 100 and the related least square estimators β 1 ( 0 ) and β 2 ( 0 ) based on the first 100 observations and later 100 observations respectively are used as initial values to implement the EM algorithm with aim to obtain the posterior mode θ 0 = ( β 1 0 , β 2 0 , k 0 ) , which is then used in the IBF sampling to get an importance sampling density π ( z | y , θ 0 ) . Secondly, { z ( j ) } j = 1 J are generated from π ( z | y , θ 0 ) with J = 6000 by SIR algorithm, and finally, L = 3000 i.i.d. posterior samples { θ ( l ) } l = 1 L are generated from π ( θ | y ) . The estimation of change-point position k is denoted as k ^ ( i ) in the i-th replication, which is the mean of L posterior samples of k, and the final estimate of k is calculated as k ^ = 1 500 i = 1 500 k ^ ( i ) . For comparison purposes, the absolute deviance (AD), the root of the mean square error (RMSE), and mean absolute deviance (MAD) of the estimated change-point position k are given as
AD = | k ^ k | , MAD = 1 500 i = 1 500 | k ^ ( i ) k | ,
and
RMSE = 1 500 i = 1 500 ( k ^ ( i ) k ) 2 .
The results based on the normal method proposed by Chen [1] are also recorded.
Table 1 presents the results. It is not surprising that when the errors come from N ( 0 , 1 ) distribution, the estimators based on the normal model to detect the position of the change-point are more accurate than the ones via quantile regression method, with smaller AD, MAD and RMSE. However, in the heavy-tailed t ( 3 ) , L ( 0 , 1 ) and Cauchy cases, overwhelmingly, the estimators based on the quantile regression approach with IBF sampler are considerately superior to the ones according to the normal method. Specifically, in the extremely heavy-tailed Cauchy situation, the normal method simply uses values around 100 to estimate the position of the change-point whenever the real k is small or large; on the contrary, the performance of the quantile regression method is relatively much better. Clearly, the quantile regression method based on the IBF sampler is more robust than the usual normal method to estimate the position of regression coefficient change-point when the data are from heavy-tailed distributions.

6. American Stock Market Data Analysis

Holbert [43] conducted a Bayesian analysis of a switching linear model and analyzed the Stock Exchange data in America, which includes n = 35 monthly observations from January 1967 to November 1969. Response variable Y is the monthly dollar volume of sales (in millions) on the Boston Stock Exchange (BSE), and regressor X is the combined New York American Stock Exchange (NYAMSE). The scatter plot shows that a suspicious change-point may exist at some time position, so a regression model with a change-point may be better to fit this data than a simple linear regression model. Chen [1] analyzed this data using the normal change-point model and detected a possible position of change-point, which is k = 23 with minimal SIC=358.18. Therefore, Chen [1] believed the regression model change starts at the time point 24, December 1968. Osorio and Galea [8] studied this data based on the Student-t model with the SIC technique, finding that k = 9 for t ( 1 ) and k = 23 when the degree of freedom is bigger than 4. Osorio and Galea [8] finally drew the same conclusion as Chen [1] that the regression model change starts at time point 24. Yang [12] detected change-point position 9 based on the Laplace regression model with a minimal SIC=353.83. Carefully examining the scatter plot, it can be seen that point 22 at the top-right is far away from other points, and y 22 is much bigger than others; thus, point 22 appears as an outlier. We consider the following robust QR change-point model
y i = β 10 + β 11 x i + ϵ i , i = 1 , , k , y i = β 20 + β 21 x i + ϵ i , i = k + 1 , , n . ε i ALD ( 0 , 1 , q ) , i = 1 , 2 , , n .
Using the above QR change-point model with 3000 generated IBF samples, we find out that the change-point possibly occurs at k = 9 for quantile level q inside interval [0.45, 0.78], and k = 23 for q outside this interval.
Figure 1 presents the scatter plots and estimated regression lines based on the normal model and QR change-point model with five quantile levels. Figure 1a presents the result by the normal model with detected change-point position 23, clearly point 22 has a high effect on the estimation, for that point 22 increases the slope of the black line by dragging this line towards itself. Figure 1b,c,f show the estimated quantile regression lines with relatively low level q = 0.2 , q = 0.4 and high level q = 0.8 respectively, and the estimated change-points are all 23, but the recognized regression lines can not effectively reflect the whole dependent relationship between BSE and NYAMSE. Figure 1d,e gives the estimated median and level q = 0.6 regression lines with change-point 9, which are robust to outlier 22 and reasonable in describing the relationship between the response and regressor. Therefore, we consider time point 9 as the position of the change-point and believe that the regression model change starts at time point 10, which is October of 1967. The median and q = 0.6 quantile regression models with change-point 9 are both reasonable. Take q = 0.6 for example, in Figure 1e, the estimated regression coefficients (black solid line) are β ^ 10 = 17.93 , β ^ 11 = 0.52 for observations 1–9, and β ^ 20 = 32.801 , β ^ 21 = 1.213 for observations 10–35 (red dashed line). This example illustrates that the quantile regression model with the IBF sampler effectively detects the position of possible change-point. As a by-product, the density histograms of the posterior samples generated by the IBF sampler from the observed posterior distribution of parameters are presented in Figure 2, which shows us how the parameters are distributed given the observed data graphically.

7. Discussion

In this paper, we proposed a change-point detection procedure based on a quantile regression model and developed a non-iterative sampling algorithm to estimate the position of the change-point. We investigated the performance of the procedure through simulations under different error distributions, finding that estimators are more accurate than ones based on the normal method when the errors follow heavy-tailed distributions, especially when the true position of the change-point is in the head part or in the tail part of the data. Finally, we applied our method to the Holbert data and detected a change-point different from the result based on the normal method. The procedure combines the robustness of quantile regression and the computational effectiveness of the non-iterative IBF sampler, so the estimations are more credible and effective than the ones based on the normal and likelihood methods. All work is done using the free R package version 4.2.2 [44]. The current work can be extended to multiple change-point problems, and further research is needed in the future.

Funding

This research was funded by the National Science Foundation of Shandong province of China under grant ZR2019MA026.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data “Holbert” used in this paper is available from the paper authored by Chen. [1].

Acknowledgments

The author would like to express his gratitude to Chen. [1] for providing the “Holbert” data.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Chen, J. Testing for a change point in linear regression model. Commun. Stat.-Theory Methods 1998, 27, 2481–2493. [Google Scholar] [CrossRef]
  2. Kim, H.J.; Fay, M.P.; Feuer, E.J.; Midthune, D.N. Permutation tests for change point regression with applications to cancer rates. Stat. Med. 2000, 19, 335–351. [Google Scholar] [CrossRef]
  3. Chen, J.; Wang, Y.P. A Statistical change point model approach for the detection of DNA copy number variations in array CGH data. IEEE/ACM Trans. Comput. Biol. Bioinform. 2009, 6, 529–541. [Google Scholar] [PubMed]
  4. Acitas, S.; Senoglu, B. Robust change point estimation in two-phase linear regression models: An application to metabolic pathway data. J. Comput. Appl. Math. 2020, 363, 337–349. [Google Scholar]
  5. Reeves, J.; Chen, J.; Wang, X.L.; Lund, R.; Lu, Q.Q. A review and comparison of changepoint detection techniques for climate data. J. Appl. Meteorol. Clim. 2007, 46, 900–915. [Google Scholar] [CrossRef]
  6. Chen, J.; Gupta, A.K. Change point analysis of a Gaussian model. Stat. Pap. 1999, 40, 323–333. [Google Scholar] [CrossRef]
  7. Chen, C.W.S.; Chan, C.S.K.; Gerlach, R.; Hsieh, W.Y.L. A comparison of estimators for regression models with change points. Stat. Comput. 2011, 21, 359–414. [Google Scholar] [CrossRef]
  8. Osorio, F.; Galea, M. Detection of a change-point in Student-t linear regression models. Stat. Pap. 2006, 47, 31–48. [Google Scholar] [CrossRef]
  9. Lin, J.G.; Chen, J.; Li, Y. Bayesian analysis of Student t linear regression with unknown change-point and application to stock data analysis. Comput. Econ. 2012, 40, 203–217. [Google Scholar]
  10. Kang, S.M.; Liu, G.Y.; Qi, H.; Wang, M. Bayesian variance changepoint detection in linear models with symmetric heavy-tailed errors. Comput. Econ. 2018, 52, 459–477. [Google Scholar]
  11. Lu, K.P.; Chang, S.T. Robust algorithms for change-point regressions using the t-distribution. Mathematics 2021, 9, 2394. [Google Scholar] [CrossRef]
  12. Yang, F.K. Robust mean change-point detecting through Laplace linear regression using EM algorithm. J. Appl. Math. 2014, 2014, 1–9. [Google Scholar]
  13. Alin, A.; Beyaztas, U.; Martin, M.A. Robust change point detection for linear regression models. Stat. Interface 2019, 12, 203–213. [Google Scholar] [CrossRef]
  14. Koenker, R.; Bassett, G. Regression quantiles. Econometrica 1978, 46, 33–50. [Google Scholar]
  15. Koenker, R. Quantile Regression, 1st ed.; Cambridge University press: Cambridge, MA, USA, 2005. [Google Scholar]
  16. Qu, Z.J. Testing for structural change in regression quantiles. J. Econom. 2008, 146, 170–184. [Google Scholar]
  17. Zhang, L.W.; Wang, H.J.; Zhu, Z.Y. Testing for change points due to a covariate threshold in quantile regression. Stat. Sin. 2014, 24, 1859–1877. [Google Scholar] [CrossRef]
  18. Gabriela, C. Estimation in a change-point non linear quantile. Commun. Stat.-Theory Methods 2017, 46, 6017–6034. [Google Scholar]
  19. Guo, J.; He, Y.H. Bayesian analysis and application of quantile auto-regression change-point model. Stat. Deci. 2017, 23, 14–18. [Google Scholar]
  20. Tan, M.; Tian, G.L.; Ng, K. A non-iterative sampling algorithm for computing posteriors in the structure of em-type algorithms. Stat. Sin. 2003, 43, 2162–2172. [Google Scholar]
  21. Tan, M.; Tian, G.L.; Ng, K. Bayesian Missing Data Problems: EM, Data Augmentation and Noniterative Computation, 1st ed.; Chapman&Hall/CRC: New York, NY, USA, 2010. [Google Scholar]
  22. Yang, F.K.; Yuan, H.J. A non-iterative posterior sampling algorithm for linear quantile regression model. Commun. Stat. Simul. Comput. 2017, 46, 5861–5878. [Google Scholar]
  23. Yang, F.K.; Yuan, H.J. A non-iterative Bayesian sampling algorithm for linear regression models with scale mixtures of normal distributions. Comput. Econ. 2017, 45, 579–597. [Google Scholar] [CrossRef]
  24. Tsionas, M.G. A non-iterative (trivial) method for posterior inference in stochastic volatility models. Stat. Probab. Lett. 2017, 126, 83–87. [Google Scholar] [CrossRef]
  25. Tian, G.L.; Ng, K.W.; Li, K.C.; Tan, M. Non-iterative sampling-based Bayesian methods for identifying changepoints in the sequence of cases of Haemolytic uraemic syndrome. Comput. Stat. Data Anal. 2009, 53, 3314–3323. [Google Scholar] [CrossRef] [PubMed]
  26. Yang, F.K.; Yuan, H.J. A non-iterative sampling algorithm for detecting the position of change-point in normal mean. Stat. Deci. 2016, 8, 16–19. [Google Scholar]
  27. Yang, F.K.; Yuan, H.J. Fast non-iterative sampling algorithm for change-point estimation of regression coefficients. Stat. Deci. 2017, 24, 10–13. [Google Scholar]
  28. Koenker, R.; Park, B.J. A interior point algorithm for nonlinear quantile regression. J. Econ. 1996, 71, 265–283. [Google Scholar] [CrossRef]
  29. Hunter, D.R.; Lange, K. Quantile regression via MM algorithm. J. Comput. Graph. Stat. 2000, 9, 60–77. [Google Scholar]
  30. Koenker, R.; Machado, J. Goodness of fit and related inference processes for quantile regression. J. Am. Stat. Assoc. 1999, 94, 1296–1309. [Google Scholar]
  31. Yu, K.; Moyeed, R. Bayesian quantile regression. Stat. Probab. Lett. 2001, 54, 437–447. [Google Scholar] [CrossRef]
  32. Reed, C.; Yu, K. A Partially Collapsed Gibbs Sampler for Bayesian Quantile Regression; Technical Report; Brunel University: London, UK, 2009. [Google Scholar]
  33. Kozumi, H.; Kobayashi, G. Gibbs sampling methods for Bayesian quantile regression. J. Stat. Comput. Simul. 2011, 81, 1565–1578. [Google Scholar] [CrossRef]
  34. Zhou, Y.H.; Ni, Z.X.; Yong, L. Quantile regression via the EM algorithm. Commun. Stat. Simul. Comput. 2014, 43, 2162–2172. [Google Scholar] [CrossRef]
  35. Tian, Y.Z.; Tian, M.Z.; Zhu, Q.Q. Linear quantile regression based on EM algorithm. Commun. Stat.-Theory Methods 2014, 43, 3464–3484. [Google Scholar] [CrossRef]
  36. Luo, Y.X.; Lian, H.; Tian, M.Z. Bayesian quantile regression for longitudinal data models. J. Stat. Comput. Simul. 2012, 82, 1635–1649. [Google Scholar] [CrossRef]
  37. Alhamzawi, R.; Yu, K. Variable selection in quantile regression via Gibbs sampling. J. Appl. Stat. 2012, 39, 799–813. [Google Scholar]
  38. Ji, Y.G.; Lin, N.; Zhang, B.X. Model selection in binary and Tobit quantile regression using the Gibbs sampler. Comput. Stat. Data Anal. 2012, 56, 827–839. [Google Scholar] [CrossRef]
  39. Tian, Y.Z.; Tang, M.L.; Tian, M.Z. A class of finite mixture of quantile regressions with its applications. J. Appl. Stat. 2016, 43, 1240–1252. [Google Scholar]
  40. Gallardo, D.I.; Bourguignon, M.; Galarza, C.E.; Gómez, H.W. A parametric quantile regression model for asymmetric response variables on the real line. Symmetry 2020, 12, 1938. [Google Scholar] [CrossRef]
  41. Reyes, J.; Rojas, M.A.; Arrué, J. A new generalization of the Student’s t distribution with an application in quantile regression. Symmetry 2021, 13, 2444. [Google Scholar] [CrossRef]
  42. Barndorff-Nielsen, O.E.; Shephard, N. Non-Gaussian Ornstein-Uhlenbeck-based models and some of their uses in financial economics. J. R. Stat. Soc. B 2001, 63, 167–241. [Google Scholar]
  43. Holbert, D. A Bayesian analysis of a switching linear model. J. Econom. 1982, 19, 77–87. [Google Scholar]
  44. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2022; Available online: https://www.R-project.org/ (accessed on 31 October 2022).
Figure 1. Scatter plots and estimated regression lines by the QR and Normal models with change-point. (a): Normal result; (bf): QR results. Black solid lines: estimated regression lines for observations 1–k; red dashed lines: estimated regression lines for observations ( k + 1 ) n.
Figure 1. Scatter plots and estimated regression lines by the QR and Normal models with change-point. (a): Normal result; (bf): QR results. Black solid lines: estimated regression lines for observations 1–k; red dashed lines: estimated regression lines for observations ( k + 1 ) n.
Symmetry 15 00447 g001
Figure 2. Density histograms of the posterior samples generated by IBF sampler from the observed posterior distribution of parameters.
Figure 2. Density histograms of the posterior samples generated by IBF sampler from the observed posterior distribution of parameters.
Symmetry 15 00447 g002
Table 1. Simulation results.
Table 1. Simulation results.
ErrorkNormal Method q = 0.25 q = 0.5 q = 0.75
ADMADRMSEADMADRMSEADMADRMSEADMADRMSE
N(0,1)400.272.424.360.373.917.510.063.626.320.834.9210.66
600.452.283.691.714.529.471.514.649.481.504.238.75
800.502.424.060.124.007.160.174.237.680.955.2412.98
1000.241.873.120.103.888.100.043.868.080.083.987.84
1200.092.443.920.153.485.350.183.244.980.083.705.95
1400.132.685.350.564.329.561.344.8510.521.544.9710.65
1600.351.993.250.324.668.660.524.687.980.354.467.98
t(3)408.9214.3635.266.1911.2329.045.8610.4627.695.19.8925.68
600.6110.2720.410.687.8814.080.507.7414.540.248.0814.85
800.549.8821.950.686.8416.310.815.019.110.94.969.06
1001.647.518.8315.511.080.356.2713.830.775.2511.47
1203.399.9923.121.017.0716.3916.1815.061.085.9815.01
1404.619.7123.680.254.998.380.255.2910.220.595.2110.11
1604.9511.2925.361.376.2711.631.426.3512.801.586.6513.67
t(1)4054.870.688.412.8027.7753.8813.3428.8055.0812.628.5854.76
6038.8362.0376.4111.0029.4954.0011.4329.5454.0211.0529.6153.60
8014.0855.3364.970.4322.1542.450.4521.5241.431.3821.4441.26
1008.8656.2863.541.1617.7835.791.4017.6035.871.3317.1334.79
12021.559.3369.334.4017.1036.494.3718.8138.295.0918.8238.75
14038.7663.0676.9412.6824.0348.0711.5523.6847.9811.5523.6447.36
16070.2780.998.1612.6430.2057.5712.5429.7257.3511.5129.0556.14
L(0,1)402.426.0216.110.814.8312.170.263.66.590.64.146.95
601.545.311.100.344.37.500.344.377.560.064.287.14
800.794.459.800.453.947.140.113.896.400.033.986.54
1000.885.0410.570.263.897.180.163.977.400.094.007.06
1200.014.217.420.124.127.020.414.227.280.074.026.82
1401.695.7113.421.035.409.850.615.149.050.975.148.96
1602.486.5617.972.326.9418.732.136.5716.801.436.4416.29
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, F. Robust Procedure for Change-Point Estimation Using Quantile Regression Model with Asymmetric Laplace Distribution. Symmetry 2023, 15, 447. https://0-doi-org.brum.beds.ac.uk/10.3390/sym15020447

AMA Style

Yang F. Robust Procedure for Change-Point Estimation Using Quantile Regression Model with Asymmetric Laplace Distribution. Symmetry. 2023; 15(2):447. https://0-doi-org.brum.beds.ac.uk/10.3390/sym15020447

Chicago/Turabian Style

Yang, Fengkai. 2023. "Robust Procedure for Change-Point Estimation Using Quantile Regression Model with Asymmetric Laplace Distribution" Symmetry 15, no. 2: 447. https://0-doi-org.brum.beds.ac.uk/10.3390/sym15020447

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop