Next Article in Journal
Evaluating Approximate Point Forecasting of Count Processes
Previous Article in Journal
A Semi-Parametric Approach to the Oaxaca–Blinder Decomposition with Continuous Group Variable and Self-Selection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bayesian Analysis of Coefficient Instability in Dynamic Regressions

Directorate General for Economics, Statistics and Research, Banca d’Italia, Via Nazionale 91, 00184 Roma, Italy
*
Author to whom correspondence should be addressed.
Submission received: 2 May 2019 / Revised: 21 June 2019 / Accepted: 25 June 2019 / Published: 28 June 2019

Abstract

:
This paper deals with instability in regression coefficients. We propose a Bayesian regression model with time-varying coefficients (TVC) that allows to jointly estimate the degree of instability and the time-path of the coefficients. Thanks to the computational tractability of the model and to the fact that it is fully automatic, we are able to run Monte Carlo experiments and analyze its finite-sample properties. We find that the estimation precision and the forecasting accuracy of the TVC model compare favorably to those of other methods commonly employed to deal with parameter instability. A distinguishing feature of the TVC model is its robustness to mis-specification: Its performance is also satisfactory when regression coefficients are stable or when they experience discrete structural breaks. As a demonstrative application, we used our TVC model to estimate the exposures of S&P 500 stocks to market-wide risk factors: We found that a vast majority of stocks had time-varying exposures and the TVC model helped to better forecast these exposures.
JEL Classification:
C11; C12; C22

There is widespread agreement that instability in regression coefficients represents a major challenge in empirical economics. In fact, many equilibrium relationships between economic variables are found to be unstable through time (e.g., Stock and Watson 1996).
An increasingly popular approach to this problem is to specify regression models with time-varying coefficients (TVC) and estimate the path of their variation (see e.g., Doan et al. 1984; Cogley and Sargent 2001; Cogley et al. 2010; D’Agostino et al. 2013; Chan 2017)1. Although the estimation of TVC models has been facilitated by advancements in Markov Chain Monte Carlo (MCMC) methods (e.g., Carter and Kohn 1994 and Chib and Greenberg 1995, and, for recent applications, Mumtaz and Theodoridis 2018; Casarin et al. 2019; Pacifico 2019), it often remains a complex task that requires a careful specification of priors and relies on computationally intensive numerical techniques. For a review and a discussion of the MCMC approach to the estimation of TVC models, see Petrova (2019).
In this paper, we propose a Bayesian TVC model that has low computational requirements and allows one to compute analytically the posterior probability that the regression is stable, the estimates of the regression coefficients, and several other quantities of interest. Furthermore, it requires minimal input from the econometrician, in the sense that priors are specified automatically: In particular, the only inputs required from the econometrician are regressors and regressands, as in ordinary least squares (OLS)regressions with constant coefficients.
Thanks to the computational tractability of our TVC model, we are able to provide a Monte Carlo study of its finite sample properties (see Kastner 2019 for another recent Monte Carlo study of a time-varying parameter model).
The main goal of the Monte Carlo experiments is to address the concerns of an applied econometrician who suspects that the coefficients of a regression might be unstable, does not know what form of instability to expect, and needs to decide what estimation strategy to adopt.
The first concern we address is loss of efficiency under the null of stability. Suppose that one’s data has indeed been generated by a regression with constant coefficients; how much does the econometrician lose, in terms of estimation precision and forecasting accuracy, when they estimate the regression using the TVC model in place of OLS? Our results suggest that the losses from using the TVC model are generally quite small and are comparable to the losses found from using frequentist breakpoint detection procedures2, such as Bai and Perron’s (1998, 2003) sequential procedure and its model-averaging variant (Pesaran and Timmerman 2007).
Another concern is the robustness to mis-specification. Suppose one’s data has been generated by a regression with few discrete structural breaks; how much does the econometrician lose from using the TVC model instead of standard frequentist procedures for breakpoint detection? The evidence from our Monte Carlo experiments indicates that as well as in this case, the estimation precision and the forecasting accuracy of the TVC model are comparable to those of standard frequentist procedures, the performance of the TVC model being slightly superior or slightly inferior depending on the sample size and on the design of the Monte Carlo experiments.
Finally, a third concern is efficiency: Even in the presence of frequently changing coefficients, does the TVC model provide better estimation precision and forecasting performance than other, possibly mis-specified, models? We found that it generally does and that this gain in efficiency can be quite large.
The Monte Carlo study is also complemented by a brief demonstration of how the TVC model can be applied to a real-world empirical problem. We considered a regression commonly employed to estimate how stock returns are related to market-wide risk factors. We found that the coefficients of this regression are unstable, with a high probability for a vast majority of the stocks included in the S&P 500 index (for applications of time-varying parameter models to stock returns, see also Bitto and Frühwirth-Schnatter 2019 and Kastner 2019). We also found that the TVC model helps to better predict the exposures of these stocks to the risk factors.
Our model belongs to the family of Class I multi-process dynamic linear models defined by West and Harrison (1997). In our specification, there is a single mixing parameter that takes on finitely many values between 0 and 1. The parameter measures the stability of regression coefficients: If it equals 0, then the regression is stable (coefficients are constant); the closer it is to 1, the more unstable the coefficients are. We propose two measures of stability that can be derived analytically from the posterior distribution of the mixing parameter, one based on credible intervals and one based on posterior odds ratios. We analyzed the performance of a simple decision rule based on these measures of stability: Use OLS if they do not provide enough evidence of instability, otherwise use TVC. We found that such a rule performs well across different scenarios, leading to very small losses under the null of stability and being still able to produce satisfactory results when coefficients are indeed unstable.
Some features of our model are borrowed from existing TVC models (in particular Doan et al. 1984; Stock and Watson 1996; Cogley and Sargent 2001). Furthermore, there are some similarities to the Dynamic Bayesian Model Averaging methods (DBMA; Hoeting et al. 1999; Raftery et al. 2010), which we discuss in Section 2. Other features of our model are completely novel. First of all, we propose an extension of Zellner’s (1986) g-prior to dynamic linear models. Thanks to this extension, posterior probabilities and coefficient estimates are invariant to rescalings of the regressors3: This property is essential for obtaining a completely automatic specification of priors. Another novelty of the model is the use of an invariant geometrically-spaced support for the prior distribution of the mixing parameter. We argue that this characteristic of the prior allows the model to capture both very low and very high degrees of coefficients instability, while retaining a considerable parsimony. Our modeling choices have two main practical consequences: (1) The priors are specified in a completely automatic way so that the regressors and regressands are the only input required from the final user4, and (2) the computational burden of the model is minimum because analytical estimators are available both for the regression coefficients and for their degree of instability.
The paper is organized as follows: Section 1 introduces the model, Section 2 describes the specification of priors, Section 3 introduces the two measures of (in)stability, Section 4 reports the results of the Monte Carlo experiments, Section 5 contains the empirical application, Section 6 concludes the discussion, and the Appendix A contains proofs and other technical details.

1. The Bayesian Model

We consider a dynamic linear model (according to West and Harrison 1997) with time-varying regression coefficients:
y t = x t β t + v t
where x t is a 1 × k vector of observable explanatory variables, β t is a k × 1 vector of unobservable regression coefficients, and v t is an i.i.d. disturbance with normal distribution having zero mean and variance V. Time is indexed by t and goes from 1 to T (T is the last observation in the sample).
The vector of coefficients β t is assumed to evolve according to the following equation:
β t = β t 1 + w t
where w t is an i.i.d. k × 1 vector of disturbances having a multivariate normal distribution with mean of zero and covariance matrix W. w t is also contemporaneously and serially independent of v t . The random walk hypothesis in Equation (2), also adopted by Cogley and Sargent (2001) and Stock and Watson (1996), implies that changes in regression coefficients happen in an unpredictable fashion.

1.1. Notation

Let information available at time t be denoted by D t . D t is defined recursively by:
D t = D t 1 y t , x t
and D 0 contains prior information on the parameters of the model (to be specified below).
We denote by z D t the distribution of a random vector z, given information at time t and by p z D t its probability density (or mass) function.
If a random vector z has a multivariate normal distribution with mean m and covariance matrix S, given D t , we write:
z D t N m , S
If a random vector z has a multivariate Student’s t distribution with mean m, scale matrix S, and n degrees of freedom, we write:
z D t T m , S , n
and its density is parameterized as follows:
p z D t n + z m S 1 z m k + n 2
If z has a Gamma distribution with parameters V and n, we write:
z D t G V , n
and its density is parameterized as follows:
p z D t = V n / 2 n / 2 z n / 2 1 exp V n z / 2 Γ n / 2
Finally, define W = V 1 W and denote by X the design matrix,
X = x 1 x T

1.2. Structure of Prior Information and Updating

In this subsection we state the main assumptions on the structure of prior information and derive the formulae for updating the priors analytically.
The first set of assumptions regards β 1 , the vector of regression coefficients at time t = 1 , and V, the variance of the regression disturbances. We impose on β 1 and V a conjugate normal-inverse Gamma prior, i.e.,
  • β 1 has a multivariate normal distribution conditional on V, with known mean β ^ 1 , 0 , and covariance equal to V · F β , 1 , 0 where F β , 1 , 0 is a known matrix;
  • The reciprocal of V has a Gamma distribution, with known parameters V ^ 0 and n 0 .
The second set of assumptions regards W. We assume that W is proportional to the prior variance of β 1 . In particular, we assume that:
W = V λ θ F β , 1 , 0
where θ is a random variable with finite support R θ = θ 1 , , θ q 0 , 1 , θ 1 = 0 , and λ θ is strictly increasing in θ , such that λ θ 1 = 0 . Hence, we assume that the innovations w t to the regression coefficients have a variance proportional to the prior variance of the regression coefficients at time t = 1 . The assumption that W F β , 1 , 0 is made also by Doan et al. (1984) in their seminal paper on TVC models5.
The coefficient of proportionality λ θ is an (unknown) random variable whose posterior distribution can be computed analytically (see below). When θ = θ 1 (and λ = 0 ), the variance of w t is zero and regression coefficients are stable. On the contrary, when θ θ 1 (and λ > 0 ), w t has a positive variance and the regression coefficients are unstable (i.e., they change through time). The higher θ is, the greater the variance of w t is, and the more unstable the regression coefficients are.
The assumptions on the priors and the initial information are summarized as follows:
Summary 1.
The priors on the unknown parameters are:
β 1 D 0 , V , θ N β ^ 1 , 0 , V F β , 1 , 0 1 / V D 0 , θ G V ^ 0 , n 0 p θ i D 0 = p 0 , i , i = 1 , , q
and the initial information set is:
D 0 = β ^ 1 , 0 , F β , 1 , 0 , V ^ 0 , n 0 , p 0 , 1 , , p 0 , q
Given the above assumptions, the posterior distributions of the parameters of the regression can be calculated as follows:
Proposition 1.
Let priors and initial information be as in Summary 1. Let p t , i = p θ = θ i D t . Then:
p β t D t 1 = i = 1 q p β t θ = θ i , D t 1 p t 1 , i p y t D t 1 , x t = i = 1 q p y t θ = θ i , D t 1 , x t p t 1 , i p 1 / V D t 1 = i = 1 q p 1 / V θ = θ i , D t 1 p t 1 , i p β t D t = i = 1 q p β t θ = θ i , D t p t , i
where,
β t θ = θ i , D t 1 T β ^ t , t 1 , i , V ^ t 1 , i F β , t , t 1 , i , n t 1 , i y t θ = θ i , D t 1 , x t T y ^ t , t 1 , i , V ^ t 1 , i F y , t , t 1 , i , n t 1 , i 1 / V θ = θ i , D t 1 G V ^ t 1 , i , n t 1 , i β t θ = θ i , D t T β ^ t , t , i , V ^ t , i F β , t , t , i , n t , i
The parameters of the above distributions are obtained recursively as:
β ^ t , t 1 , i = β ^ t 1 , t 1 , i F β , t , t 1 , i = F β , t 1 , t 1 , i + λ θ i F β , 1 , 0 y ^ t , t 1 , i = x t β ^ t , t 1 , i F y , t , t 1 , i = 1 + x t F β , t , t 1 , i x t e t , i = y t y ^ t , t 1 , i P t , i = F β , t , t 1 , i x t / F y , t , t 1 , i β ^ t , t , i = β ^ t , t 1 , i + P t , i e t , i F β , t , t , i = F β , t , t 1 , i P t , i P t , i F y , t , t 1 , i n t , i = n t 1 , i + 1 V ^ t , i = 1 n t , i n t 1 , i V ^ t 1 , i + e t , i 2 F y , t , t 1 , i
starting from the initial conditions β ^ 1 , 0 , i = β ^ 1 , 0 , F β , 1 , 0 , i = F β , 1 , 0 , V ^ 0 , i = V ^ 0 , and n 0 , i = n 0 . The mixing probabilities are obtained recursively as:
p t , i = p t 1 , i p y t θ = θ i , D t 1 , x t j = 1 q p t 1 , j p y t θ = θ j , D t 1 , x t
starting from the prior probabilities p 0 , 1 , , p 0 , q .
The updated mixing probabilities in the above proposition can be interpreted as posterior model probabilities, where a model is a TVC regression with fixed θ . Hence, for example, p T , 1 is the posterior probability of the regression model with stable coefficients ( θ = 0 ). A crucial property of the framework we propose is that posterior model probabilities are known analytically: They can be computed exactly, without resorting to simulations.
We note that the above recursive equations bear some similarities to those found in Dynamic Bayesian Model Averaging (DBMA; Hoeting et al. 1999; Raftery et al. 2010). The main difference is in the update of the covariance matrix of coefficients. In our model we have:
F β , t , t 1 , i = F β , t 1 , t 1 , i + λ θ i F β , 1 , 0
while in DBMA, there is:
F β , t , t 1 , i = λ 1 F β , t 1 , t 1 , i
where λ is a forgetting factor that was exogenously set at λ = 0.99 by Raftery et al. (2010). Furthermore, while in DBMA a forgetting factor is also applied to model probabilities p t , i , the recursions for mixing probabilities in our model have no forgetting factor because we assume that θ stays constant through time. Applying a forgetting factor as in DBMA to the mixing probabilities would be a straightforward extension of our model. Finally, while Raftery et al. (2010) use a recursive moment estimator for V ^ t , i , we derive it in a fully Bayesian way.
In the above proposition, the priors on the regression coefficients β t in a generic time period t are updated using information received up to that same time t only. However, after observing the whole sample (up to time T), one might want to revise the priors on the regression coefficients β t in previous time periods ( t < T ), using the information subsequently received. This revision (usually referred to as smoothing) can be accomplished using the results of the following proposition:
Proposition 2.
Let priors and initial information be as in Summary 1. Then:
p β T τ D T = i = 1 q p β T τ θ = θ i , D T p T , i
where,
β T τ θ = θ i , D T T β ^ T τ , T , i , V ^ T , i F β , T τ , T , i , n T , i
The mixing probabilities p T , i and the parameters V ^ T , i , n T , i are obtained from the recursions in Proposition 1 while the parameters β ^ T τ , T , i and F β , T τ , T , i are obtained from the following backward recursions:
Q T τ , i = F β , T τ , T τ , i F β , T τ + 1 , T τ , i 1 β ^ T τ , T , i = β ^ T τ , T τ , i + Q T τ , i β ^ T τ + 1 , T , i β ^ T τ + 1 , T τ , i F β , T τ , T , i = F β , T τ , T τ , i + Q T τ , i F β , T τ + 1 , T , i F β , T τ + 1 , T τ , i Q T τ , i
starting from τ = 1 and taking as the final conditions the values β ^ T 1 , T 1 , i , β ^ T , T , i , β ^ T , T 1 , i , F β , T 1 , T 1 , i , F β , T , T , i , and F β , T , T 1 , i obtained from the recursions in Proposition 1.
Other important quantities of interest are known analytically, as shown by the following lemma:
Lemma 1.
The following quantities are also known analytically:
E β t D s = i = 1 q p s , i E β t θ = θ i , D s E y t D s = i = 1 q p s , i E y t θ = θ i , D s Var β t D s = i = 1 q p s , i Var β t θ = θ i , D s + i = 1 q p s , i E β t θ = θ i , D s E β t θ = θ i , D s i = 1 q p s , i E β t θ = θ i , D s i = 1 q p s , i E β t θ = θ i , D s Var y t D s = i = 1 q p s , i Var y t θ = θ i , D s + i = 1 q p s , i E y t θ = θ i , D s 2 i = 1 q p s , i E y t θ = θ i , D s 2
where E β t θ = θ i , D s , E y t D s , Var β t D s , and Var y t D s are calculated for each θ i as in Propositions 1 and 2.
Thus, parameter estimates ( E β t D s ) and predictions ( E y t D s ) in any time period can be computed analytically and their variances are known in closed form. The probability distributions of β t and y t in a certain time period given information D s are mixtures of Student’s t distributions. Their quantiles are not known analytically, but they are easy to simulate by Monte Carlo methods. For example, if the distribution of β T conditional on D T is the object of interest, one can set up a Monte Carlo experiment where each simulation is conducted in two steps: (1) Extract z from a uniform distribution on 0 , 1 ; find k such that k = arg min k : i = 1 k p T , i z , and (2) given k , extract β T from the Student’s t distribution β T θ = θ k , D T , which is given by Propositions 1 and 2. The empirical distribution of the Monte Carlo simulations of β T thus obtained is an estimate of the distribution of β T conditional on D T .

2. The Specification of Priors

Our specification of priors aims to be:
  • Objective, in the sense that it does not require elicitation of subjective priors;
  • Fully automatic, in the sense that the model necessitates no inputs from the econometrician other than regressors and regressands, as in OLS regressions with constant coefficients.
The above goals are pursued by extending Zellner’s (1986) g-prior to TVC models and by parameterizing λ θ in such a way that the support of θ is invariant (it does not need to be specified on a case-by-case basis).

2.1. The Prior Mean and Variance of the Coefficients

We use a version of Zellner’s (1986) g-prior for the prior distribution of the regression coefficients at time t = 1 . The prior mean is zero, corresponding to a prior belief of no predictability:
β ^ 1 , 0 = 0
while the prior covariance matrix is proportional to X X 1 :
F β , 1 , 0 = g X X 1
where g is a coefficient of proportionality.
Zellner’s (1986) g-prior is widely used in model selection and model averaging problems similar to ours (we have a range of regression models featuring different degrees of instability) because it greatly reduces the sensitivity of posterior model probabilities to the specification of prior distributions (Fernandez et al. 2001), thus helping to keep the analysis as objective as possible. Furthermore, Zellner’s (1986) g-prior has a straightforward interpretation: It can be interpreted as information provided by a conceptual sample having the same design matrix X as the current sample (Zellner 1986; George and McCulloch 1997; Smith and Kohn 1996).
To keep the prior relatively uninformative, we follow Kass and Wasserman (1995) and choose g = T (see also Shively et al. 1999). Thus, the amount of prior information (in the Fisher sense) about the coefficients is equal to the average amount of information contained in one observation from the sample.
Given the assumption that W = λ θ F β , 1 , 0 (see previous section), Zellner’s prior (4) implies that the covariance matrix of w t is also proportional to X X 1 :
W X X 1
The above proportionality has also been assumed in a TVC model by Stock and Watson (1996)6, who borrow it from Nyblom (1989). A similar hypothesis is also adopted by Cogley and Sargent (2001)7.
Equations (4) and (3), together with Equation (5), imply that all the coefficients β t have zero prior mean and covariance proportional to X X 1 , given D 0 :
E β t D 0 , V , θ = 0 , t = 1 , T , θ Var β t D 0 , V , θ = T + t 1 λ θ V X X 1 , t = 1 , , T , θ
As a consequence, the model has the desirable property that the posterior model probabilities are scale invariant in the covariates (Nyblom 1989).

2.2. The Variance Parameters V ^ 0 and n 0

In objective Bayesian analyses, the prior usually assigned to V in conjunction with Zellner’s (1986) g-prior (e.g., Liang et al. 2008) is the improper prior:
p V D 0 , θ V 1
With this choice, the updating equations in Proposition 1 would have to be replaced with a different set of updating equations until the first non-zero observation of y t (West and Harrison 1997). Furthermore, the updating of posterior probabilities would be slightly more complicated. To avoid the subtleties involved in using an improper prior, we adopt a simpler procedure that yields almost identical results in reasonably sized samples. We use the first observation in the sample (denote it by y 0 ) to form the prior on V:
V ^ 0 = y 0 2 n 0 = 1
We then discard the observation and start the updating in Proposition 1 from the subsequent observation. If the first observation is zero ( y 0 = 0 ) we discard it and use the next to form the prior (or repeat until we find the next non-zero observation).

2.3. The Mixing Parameter θ and the Prior Mixing Probabilities p 0 , i

When relaxing the assumption that the matrix W is known, we have assumed that:
W = λ θ F β , 1 , 0
where θ is a random variable with finite support R θ = θ 1 , , θ q 0 , 1 , θ 1 = 0 and λ θ is strictly increasing in θ , such that λ θ 1 = 0 . We now propose a specification of the function λ θ that satisfies the above requirements and allows for an intuitive interpretation of the parameter θ , while also facilitating the specification of a prior distribution for θ .
First, note that:
y t = x t β t 1 + x t w t + v t
Hence, given λ and the initial information D 0 , the variance generated by innovations at time t is:
Var x t w t + v t D 0 , λ = λ V ^ 0 x t F β , 1 , 0 x t + V ^ 0
We assume that θ is the fraction of this variance generated on average by innovations to the coefficients:
θ = 1 T t = 1 T Var x t w t D 0 , λ 1 T t = 1 T Var x t w t + v t D 0 , λ
When θ satisfies the above property, it is immediate to prove that:
λ θ = θ ω θ ω
where,
ω = 1 T t = 1 T x t F β , 1 , 0 x t
The function λ θ is strictly increasing in θ , such that λ 0 = 0 , as required. Hence, when θ = 0 , the regression has stable coefficients. Furthermore, by an appropriate choice of θ , any degree of coefficient instability can be reproduced (when θ tends to 1, λ approaches infinity).
The support of θ is a geometrically spaced grid, consisting of q points:
R θ = 0 , θ max c q 2 , θ max c q 1 , , θ max c 1 , θ max
where 0 θ max < 1 and 0 < c < 1 . θ max cannot be chosen to be exactly equal to 1 (in which case λ θ max = ), but it can be set equal to any number arbitrarily close to 1.
If the grid is considered as an approximation of a finer set of points (possibly a continuum), the geometric spacing ensures that the maximum percentage round-off error is constant on all subintervals θ i , θ i + 1 ( 1 < i < q ). 1 c 2 is the constant that bounds the percentage round-off error.
Using a geometrically spaced grid is the natural choice when the order of magnitude of a parameter is unknown (e.g., Guerre and Lavergne 2005, Horowitz and Spokoiny 2001; Lepski et al. 1997, and Spokoiny 2001). In our case, it allows one to simultaneously consider both regressions that are very close to being stable and regressions that are far from being stable, without requiring too fine a grid8.
Assuming prior ignorance on the order of magnitude of θ , we assign equal probability to each point in the grid:
p θ i D 0 = q 1 , i = 1 , , q
Note that, given the above choices, the prior on θ and its support are invariant, in the sense that they do not depend on any specific characteristic of the data to be analyzed, but they depend only on the maximum percentage round-off error 1 c 2 . As a consequence, they allow the specification of priors to remain fully automatic.

3. Measures of (In)Stability

After computing the posterior distribution of θ , a researcher might naturally ask: How much evidence did the data provide against the hypothesis of stability? Here, we discuss some possible ways to answer this question.
The crudest way to evaluate instability is to look at the posterior probability that θ = 0 . The closer to 1 this probability is, the more evidence of stability we have. However, a low posterior probability that θ = 0 does not necessarily constitute overwhelming evidence of instability. It might simply be the case that the sample is not large enough to satisfactorily discriminate, a posteriori, between stable and unstable regressions: In such cases, even if the true regression is stable, unstable regressions might be assigned posterior probabilities that are only marginally lower than the probability of the stable one. Furthermore, if R θ contains a great number of points, it can happen that the posterior probability that θ = 0 is close to zero, but still much higher than the posterior probability of all the other points.
We propose two measures of stability to help circumvent the above shortcomings. The first one is based on credible intervals (e.g., Robert 2007). The second one is based on posterior odds ratios.
Define the higher posterior probability set H θ as follows:
H θ = θ i R θ : p θ = θ i D T > p θ = 0 D T
i.e., H θ contains all points of R θ having higher posterior probability than θ = 0 (remember that θ = 0 means that regression coefficients are stable). Define Π as follows:
Π = 1 θ i H θ p θ = θ i D T θ i 0 p θ = θ i D T
where we adopt the convention 0 / 0 = 0 .
When Π = 1 , θ = 0 is a mode of the posterior distribution of θ : We attach to the hypothesis of stability a posterior probability that is at least as high as the posterior probability of any alternative hypothesis of instability. On the contrary, when Π = 0 , the posterior probability assigned to the hypothesis of stability is so low that all unstable models are more likely than the stable one, a posteriori. In the intermediate cases ( 0 < Π < 1 ), Π provides a measure of how far the hypothesis of stability is from being the most likely hypothesis (the lower Π , the less likely stability is).
The second measure we propose is based on the probability of the posterior mode of θ . Define:
p = max p θ = θ i D T : θ i R θ
p is the probability of (one of) the mode(s) of the posterior distribution of θ , i.e., the probability of the most likely value(s) of θ .
Using p , we construct our second measure of stability as a posterior odds ratio:
π = p θ = 0 D T p
As with the previously proposed measure, when π = 1 , θ = 0 is a mode of the posterior distribution of θ and stability is the most likely hypothesis, a posteriori. On the contrary, the closer π is to zero, the less likely stability is, when compared with the most likely hypothesis. For example, when π = 1 / 10 , there is an unstable regression that is 10 times more likely than the stable one.
Both measures of stability ( Π and π ) can be used to make decisions. For example, one can fix a threshold τ and decide to reject the hypothesis of stability if the measure of stability is below the threshold ( Π < τ or π < τ ). In that case that Π is used, the procedure can be assimilated to a frequentist test of hypothesis, where 1 τ represents the level of confidence. Π can be interpreted as a sort of Bayesian p-value (e.g., Robert 2007): The lower Π is, the higher is the confidence with which we can reject the hypothesis of stability9. In that case that π is used, one can resort to Jeffreys’s (1961) scale to qualitatively assess the strength of the evidence against the hypothesis of stability (e.g., substantial evidence if 1 3 π < 1 10 , strong evidence if 1 10 π < 1 30 , very strong evidence if 1 30 π < 1 100 ).
In the next section we explore the consequences of using these decision rules to decide whether to estimate a regression by OLS or by TVC.

4. Monte Carlo Evidence

4.1. Performance When the Data Generating Process (DGP) Is a Stable Regression

In this subsection we present the results of a set of Monte Carlo simulations aimed at evaluating how much efficiency is lost when a stable regression is estimated with our TVC model. We compare the forecasting performance and the estimation precision of the TVC model with those of OLS and of a standard frequentist procedure used to identify breakpoints and estimate regression coefficients in the presence of structural breaks. In particular, we consider the performance of Bai and Perron’s (1998, 2003) sequential procedure, as implemented by Pesaran and Timmerman (2002, 2007).
For our Monte Carlo experiments, we adapt a design that has already been employed in the literature on parameter instability (Hansen 2000).
The design is as follows:
  • Data generating process: y t is generated according to:
    y t = ρ y t 1 + u t 1 + v t
    where y 0 = 0 , u t T 0 , 1 , 5 i.i.d., v t N 0 , 1 i.i.d., and u t and z t are serially and cross-sectionally independent;
  • Estimated equations: Two equations are estimated. In the first case, a constant and the first lags of y t and u t are included in the set of regressors; hence, the estimated model is (1), where:
    x t = 1 y t 1 u t 1
    In the second case, a constant and the first three lags of y t and u t are included in the set of regressors; hence, the estimated model is (1), where:
    x t = 1 y t 1 y t 2 y t 3 u t 1 u t 2 u t 3
  • Parameters of the design: Simulations are conducted for three different sample sizes ( T = 100 , 200 , 500 ), four different values of the autoregressive coefficient ( ρ = 0 , 0.50 , 0.80 , 0.99 ), and the two estimated equations detailed above, for a total of 24 experiments.
Each Monte Carlo experiment consists of 10,000 simulations. The loss in estimation precision is evaluated by comparing the estimate of the coefficient vector at time T (denote it by β ˜ T ) with its true value. We consider seven different estimates:
  • Model averaging (TVC-MA) estimates, where:
    β ˜ T = E β T D T = i = 1 q p T , i E β T D T , θ = θ i
  • Model selection (TVC-MS) estimates, where:
    β ˜ T = E β T D T , θ = θ j
    and
    j = arg max j p T , j
    i.e., only the model with the highest posterior probability is used to make predictions;
  • Estimates obtained from the regression model with stable coefficients when Π 0.1 and from model averaging when Π < 0.1 (denoted by TVC- Π ):
    β ˜ T = E β T D T , θ = 0   if   Π 0.1 E β T D T   if   Π < 0.1
    i.e., coefficients are estimated with the TVC model only if there is enough evidence of instability ( Π < 0.1 ); otherwise, the standard OLS regression is used. This is intended to reproduce the outcomes of a decision rule whereby the econometrician uses the TVC model only if the TVC model itself provides enough evidence that OLS is inadequate;
  • Estimates obtained from the regression model with stable coefficients when π 0.1 and from model averaging when π < 0.1 (denoted by TVC- π ):
    β ˜ T = E β T D T , θ = 0   if   π 0.1 E β T D T   if   π < 0.1
    This estimator is similar to the previous one, but π is used in place of Π to decide whether there is enough evidence of instability;
  • Estimates obtained from the regression model with stable coefficients (OLS):
    β ˜ T = E β T D T , θ = 0
  • OLS estimates obtained from Bai and Perron’s (1998, 2003) sequential10 procedure (denoted by BP), using the Schwarz information criterion (SIC) criterion to choose the number of breakpoints (Pesaran and Timmerman 2002 and 2007). If τ ˜ is the last estimated breakpoint date in the sample, then β ˜ T is the OLS estimate of β T obtained by using all the sample points from τ ˜ to T;
  • Estimates obtained from Pesaran and Timmerman’s (2007) model-averaging procedure (denoted by BP-MA): The location of the last breakpoint is estimated with Bai and Perron’s procedure (as in the point above); if τ ˜ is the last estimated breakpoint date in the sample, then:
    β ˜ T = τ = 1 τ ˜ w τ β ˜ T , τ
    where β ˜ T , τ is the OLS estimate of β T obtained using all the sample points from τ to T; w τ is a weight proportional to the inverse of the mean squared prediction error committed when using only the sample points from τ onwards to estimate the regression and predict y t ( τ + k + 1 t T ).
The Monte Carlo replications are used to estimate the mean squared error (MSE) of the coefficient estimates:
M S E j β = E β T β ˜ T 2
where || || is the Euclidean norm and j = TVC-MA, TVC-MS, TVC- Π , TVC- π , OLS, BP, BP-MA depending on which of the above methods has been used to estimate β T .
The two parameters regulating the granularity of the grid for θ are chosen as follows: q = 100 and c = 0.9 . To avoid degeneracies, rather than setting θ max = 1 (the theoretical upper bound on θ ), we chose a value that is numerically close to 1 ( θ max = 0.999 ). Thus, the roundoff error is bounded at 5 per cent and the model is able to detect both very high and very low degrees of instability (as low as θ 3 · 10 5 ).
Panel A of Table 1 reports the Monte Carlo estimates of M S E j β for the case in which x t includes only the first lags of y t and u t . Not surprisingly, the smallest MSE is in all cases achieved by the OLS estimates. There are significant differences between the case in which the autoregressive component is very persistent ( ρ = 0.99 ) and the other cases ( ρ = 0 , 0.50 , 0.80 ). In the latter cases, the TVC- π coefficient estimates are those that yield the smallest increase in MSE with respect to OLS (in most cases under 5 per cent). The performance of BP-MA is the second best, being only slightly inferior to that of TVC- π , but slightly superior to that of TVC- Π . M S E T V C M A β and M S E T V C M S β are roughly between 20 per cent and 60 per cent higher than M S E T V C O L S β , while M S E T V C B P β is on average equal to several multiples of M S E T V C O L S β . Qualitatively speaking, the loss in precision from using TVC- Π , TVC- π , and BP-MA is almost negligible, while there is a severe loss using BP and a moderate loss using TVC-MA and TVC-MS. In the case in which ρ = 0.99 , results are very different: On average, M S E T V C β (all four kinds of TVC) and M S E B P β become almost two orders of magnitude greater than M S E O L S β , while M S E B P M A β remains comparable to M S E O L S β (although there is a worsening with respect to the case of low persistence).
The unsatisfactory performance of the TVC and BP estimates in the case of high persistence can arguably be explained by an identification problem. In the unit root case, the regression generating the data is:
y t = y t 1 + u t 1 + v t
For any α < 1 , it can be rewritten as:
y t = μ t + α y t 1 + u t 1 + v t
where μ t = 1 α y t 1 is an intercept following a random walk. Furthermore, its innovations ( μ t μ t 1 ) are contemporaneously independent of the innovations v t . Therefore, if the estimated equation includes a constant and time-varying coefficients are not ruled out, it is not possible to identify whether the regression has a unit root and stable coefficients, or has a stationary autoregressive component and a time-varying intercept11. When ρ is near unity, identification is possible, but it will presumably be weak, giving rise to very imprecise estimates of the coefficients and of their degree of stability. Note that the two equivalent (and unidentified) representations above obviously yield the same one-step-ahead forecasts of y t . Therefore, if our conjecture that this weak identification problem is affecting our results is correct, we should find that the out-of-sample forecasts of y t produced by the TVC model are not as unsatisfactory as its coefficient estimates. This is exactly what we find and document in the last part of this subsection.
Panel B of Table 1 reports the Monte Carlo estimates of M S E j β for the case in which x t includes three lags of y t and u t . In the case of low persistence, the BP-MA estimates are those that achieve the smallest increase in MSE with respect to the OLS estimates (on average below 2 per cent). The performance of the TVC- π estimates is only slightly inferior (around a 3 percent increase in MSE with respect to OLS). All the other estimates (TVC-MA, TVC-MS, TVC- Π , and BP) are somewhat less efficient, but their MSEs seldom exceed those of the OLS estimates by more than 30 per cent. As far as the highly persistent case ( ρ = 0.99 ) is concerned, we again observe a degradation in the performance of the TVC and (to a lesser extent) of the BP estimates. However, the degradation is less severe than the one observed in the case of fewer regressors. Intuitively, adding more regressors (even if their coefficients are 0) helps to alleviate the identification problem discussed before, because the added regressors have stable coefficients and hence help to pin down the stable representation of the regression.
The loss in forecasting performance is evaluated using a single out-of-sample prediction for each replication. In each replication, T + 1 observations are generated, the first T are used to update the priors, the vector of regressors x T + 1 is used to predict y T + 1 , and the prediction (denote it by y ˜ T + 1 ) is compared to the actual value y T + 1 . As for coefficient estimates, we consider seven different predictions:
  • Model averaging (TVC-MA) predictions, where:
    y ˜ T + 1 = E y T + 1 D T , x T + 1 = i = 1 q p T , i E y T + 1 D T , x T + 1 , θ = θ i
  • Model selection (TVC-MS) predictions, where:
    y ˜ T + 1 = E y T + 1 D T , x T + 1 , θ = θ j
  • Predictions generated by the regression model with stable coefficients when Π 0.1 and by model averaging when Π < 0.1 (denoted by TVC- Π ):
    y ˜ T + 1 = E y T + 1 D T , x T + 1 , θ = 0   if   Π 0.1 E y T + 1 D T , x T + 1   if   Π < 0.1
  • Predictions generated by the regression model with stable coefficients when π 0.1 and by model averaging when π < 0.1 (denoted by TVC- π ):
    y ˜ T + 1 = E y T + 1 D T , x T + 1 , θ = 0   if   π 0.1 E y T + 1 D T , x T + 1   if   π < 0.1
  • Predictions generated by the regression model with stable coefficients (OLS):
    y ˜ T + 1 = E y T + 1 D T , x T + 1 , θ = 0
  • Predictions obtained from Bai and Perron’s sequential procedure (BP); if β ˜ T is the BP estimate of β T (see above), then:
    y ˜ T + 1 = x T + 1 β ˜ T
  • Predictions obtained from Pesaran and Timmerman’s (2007) model-averaging procedure (BP-MA); if β ˜ T is the BP-MA estimate of β T (see above), then:
    y ˜ T + 1 = x T + 1 β ˜ T
The Monte Carlo replications are used to estimate the mean squared error of the predictions:
M S E j y = E y T + 1 y ˜ T + 1 2
where j = TVC-MA, TVC-MS, TVC- Π , TVC- π , OLS, BP, BP-MA depending on which of the above methods has been used to forecast y T + 1 .
To improve the accuracy of our Monte Carlo estimates of M S E j y , we use the fact that:
E y T + 1 y ˜ T + 1 2 = E v T + 1 2 + E β T + 1 β ˜ T + 1 x T + 1 x T + 1 β T + 1 β ˜ T + 1
Since E v T + 1 2 is known, we use the Monte Carlo simulations to estimate only the second summand on the right hand side of the above equation.
Table 2 reports the Monte Carlo estimates of M S E j y . The variation in M S E j y across models and design parameters broadly reflects the variation in M S E j β we have discussed above. To avoid repetitions, we point out the only significant difference, which concerns the highly persistent design ( ρ = 0.99 ), while the TVC and BP estimates give rise to an M S E j β that is around two orders of magnitude higher than M S E O L S β , the part of their M S E j y attributable to estimation error ( M S E j y 1 ) compares much more favorably to its OLS counterpart, especially in the designs where x t includes three lags of y t and u t . This might be considered evidence of the identification problem mentioned above.

4.2. Performance When the DGP Is a Regression with a Discrete Structural Break

In this subsection we present the results of a set of Monte Carlo simulations aimed at understanding how our TVC model performs when regression coefficients experience a single discrete structural break. As in the previous subsection, we analyze both losses in forecasting performance and losses in estimation precision.
The Monte Carlo design is the same one employed in the previous subsection, except for the fact that the data generating process is now subject to a discrete structural break at an unknown date:
  • Data generating process: y t is generated according to:
    y t = ρ y t 1 + u t 1 + v t if t < τ y t = ρ y t 1 + 1 + b u t 1 + v t if t τ
    where y 0 = 0 , u t T 0 , 1 , 5 i.i.d., v t N 0 , 1 i.i.d., and u t and v t are serially and cross-sectionally independent; τ is the stochastic breakpoint date, extracted from a discrete uniform distribution on the set of sample dates (from 1 to T); b N 0 , 1 is the stochastic break in regression coefficients.
The estimation precision and the forecasting performance are evaluated by comparing the estimates of the coefficient vector at time T and the predictions of y T + 1 with their true values.
Panel A of Table 3 reports the Monte Carlo estimates of M S E j β for the case in which x t includes only the first lags of y t and u t . As before, we first discuss the cases in which ρ 0.99 . The OLS estimates, which have the smallest MSEs in the stable case (see previous subsections) are now those with the highest MSEs. Both the frequentist methods (BP and BP-MS) and the TVC methods (all four kinds) achieve a significant reduction in the MSEs with respect to OLS. Although TVC-MA and TVC-MS perform slightly better than TVC- Π and TVC- π , there is not a clear ranking between the former two and the two frequentist methods: Their MSEs are on average comparable, but TVC-MA and TVC-MS tend to perform better when the sample size is small ( T = 100 ), while BP and BP-MA tend to perform better when the sample size is large ( T = 200 , 500 ). This might be explained by the fact that BP and BP-MA require the estimation of a considerable number of parameters when one or more breakdates are found and these parameters are inevitably estimated with low precision when the sample size is small. In the case in which ρ = 0.99 , results are again substantially different: The MSEs of the TVC estimates (all four kinds) and of the BP estimates become much larger than the MSEs of the OLS estimates (and the BP estimates fare better than the TVC estimates), while the MSEs of the BP-MA estimates remain below those of the OLS estimates. The remarks about potential identification problems made in the previous subsections also apply to these results.
Panel B of Table 3 reports the Monte Carlo estimates of M S E j β for the case in which x t includes three lags of y t and u t . The patterns are roughly the same found in Panel A (see previous paragraph), with the relative performance of the TVC methods and the frequentist methods depending on the sample size T. The only difference worth mentioning is that when ρ = 0.99 , the increase in the MSEs is milder and the TVC-MA estimates are more precise than the BP estimates.
As far as out-of-sample forecasting performance is concerned (Table 4, Panels A and B), the patterns in the M S E j y broadly reflect the patterns in the M S E j β . Again, there is an exception to this: When ρ = 0.99 , high values of M S E j β do not translate into high values of M S E j y and as a consequence, despite the aforementioned identification problem, the BP and the four TVC forecasts are much more accurate than the OLS forecasts (and in some cases are also more accurate than the BP-MA forecasts).

4.3. Performance When the DGP Is a Regression with Frequently Changing Coefficients

In this subsection we present the results of a set of Monte Carlo simulations aimed at understanding how our TVC model performs when regression coefficients experience frequent changes.
We analyze both losses in forecasting performance and losses in estimation precision, using the same Monte Carlo design employed in the previous two subsections. The only difference is that the data is now generated by a regression whose coefficients change at every time period:
  • Data generating process: y t is generated according to:
    y t = ρ y t 1 + b t u t 1 + v t b t = b t 1 + w t
    where y 0 = 0 , b 0 = 1 , u t T 0 , 1 , 5 i.i.d., v t N 0 , 1 i.i.d., w t N 0 , V i.i.d., and u t , v t and w t are serially and cross-sectionally independent. To ease comparisons with the previous subsection, V is chosen in such a way that b T N 1 , 1 , irrespective of the sample size T:
    V = 1 T
Note that, although one coefficient of the regression is frequently changing ( b t ), the other coefficient ( ρ ) is stable, therefore the true DGP does not exactly fit any of the possible DGPs contemplated by the TVC model. We prefer to adopt this specification over a specification in which the TVC model is correctly specified, because the results obtained with the latter specification are trivial (the TVC estimates are the best possible estimates). Furthermore, controlling ρ (keeping it fixed) allows a better understanding of its effects on model performance.
Panel A of Table 5 reports the Monte Carlo estimates of M S E j β for the case in which x t includes only the first lags of y t and u t . We first summarize the results obtained when ρ 0.99 . The lowest MSEs are achieved by the TVC-MA estimates. The TVC-MS estimates are the second best (in some cases M S E T V C M S β is almost identical to M S E T V C M A β ). TVC- Π and TVC- π also have a performance comparable to that of TVC-MA (the increase in the MSEs is on average less than 5 per cent). The BP estimates are significantly less precise than the TVC estimates (their MSEs are roughly between 30 per cent and 70 per cent higher than M S E T V C M A β ). Finally, BP and BP-MA have a comparable performance when T = 100 , but BP-MA is much less precise when the sample size increases ( T = 200 , 500 ).
When ρ = 0.99 , we again observe a sharp increase in the MSEs of the TVC estimates (all four kinds) and of the BP estimates: Their MSEs become several times those of the OLS estimates. BP-MA achieves a significant reduction in MSE over OLS with larger sample sizes ( T = 200 , 500 ). Thus, with frequently changing coefficients, BP-MA seems to be the only method capable of simultaneously dealing with coefficient instability and a highly persistent lagged dependent variable.
Panel B of Table 5 reports the Monte Carlo estimates of M S E j β for the case in which x t includes three lags of y t and u t . Similar to what we found in the previous subsections, the only noticeable difference with respect to the one-lag case is that when ρ = 0.99 the increase in the MSEs is milder.
As far as out-of-sample forecasting performance is concerned (Table 6, Panels A and B), the patterns in the M S E j y broadly reflect the patterns in the M S E j β . Again, the case of ρ = 0.99 constitutes an exception. Despite their high M S E j β , the BP and the four TVC forecasts are more accurate than the OLS forecasts (and the TVC-MA and TVC- Π forecasts are also more accurate than the BP-MA forecasts).

5. Empirical Application: Estimating Common Stocks’ Exposures to Risk Factors

In this section we briefly illustrate an empirical application of our TVC model. We use the model to estimate the exposures of S&P 500 constituents to market-wide risk factors. We track the weekly returns of the S&P 500 constituents for 10 years (from January 2000 to December 2009). An uninterrupted time series of returns is available for 432 of the 500 constituents (as of December 2009). The list of constituents and their returns are downloaded from Datastream. The risk factors we consider are the Fama and French’s (1993, 1995, 1996) risk factors (excess return on the market portfolio, return on the Small Minus Big portfolio, return on the High Minus Low portfolio), downloaded from Kenneth French’s website.
The exposures to the risk factors are the coefficients β t in the regression
y t = x t β t + v t
where y t is the excess return on a stock at time t,
x t = 1 r M , t r f , t S M B t H M L t
r M , t is the return on the market portfolio at time t, r f , t is the risk-free rate of return, and S M B t and H M L t are the returns at time t on the SMB and HML portfolios respectively.
The procedures illustrated in the previous section are employed to understand whether the risk exposures β t are time-varying and whether the TVC model provides good estimates of these risk exposures.
For a vast majority of the stocks included in our sample, we find evidence that β t is indeed time-varying. θ = 0 is the posterior mode of the mixing parameter for only 11 stocks out of 432. Furthermore, P < 0.1 and π < 0.1 for 92 per cent and 81 per cent of the stocks respectively. On average, P is 0.046 and π is 0.010. Furthermore, the frequentist method provides evidence that most stocks experience instability in their risk exposures. According to the BP sequential estimates, more than 78 per cent of stocks experience at least one break in β t .
To evaluate the forecasting performance, we use the out-of-sample forecasts of y t obtained after the first 400th week. The methods used to make predictions are those described in the previous section ( j = TVC-MA, TVC-MS, TVC- Π , TVC- π , OLS, BP, BP-MA ). For each stock i and for each prediction method j, the mean squared error is computed as:
M S E i , j ¯ = 1 T T 0 t = T 0 + 1 T y t , i , j y ˜ t , i , j 2
where T 0 is the number of periods elapsed before the first out-of-sample forecast is produced, y ˜ t , i , j denotes the prediction of the excess return of the i-th stock at time t, conditional on x t , produced by method j, and y t , i , j is the corresponding realization.
In order to compare the performance of the various methods across stocks, we use the performance of OLS forecasts as a benchmark. Thus, the gain from using model j with stock i is defined as:
G A I N i , j = 1 M S E i , j ¯ M S E i , O L S ¯
i.e., G A I N i , j is the average reduction in MSE achieved by using model j instead of OLS. A positive value indicates an improvement in forecasting performance.
Table 7 reports some summary statistics of the sample distribution of G A I N i , j (to each stock i corresponds a different sample point). All the TVC methods achieve a reduction in MSE and, among the TVC methods, TVC-MA achieves the maximum average reduction (approximately 3 per cent). BP performs very poorly (it actually causes a strong increase in MSE), while the average reduction achieved by BP-MA is similar to that of TVC-MA (again, approximately 3 per cent). The four TVC models have similar sample distributions of gains, characterized by a pronounced skew to the right (several small gains and few very large gains); furthermore, all four have a more dispersed distribution than the BP-MA model.

6. Conclusions

We have proposed a Bayesian regression model with time-varying coefficients (TVC) that has low computational requirements because it allows one to derive analytically the posterior distribution of coefficients, as well as the posterior probability that they are stable.
The model is completely automatic in the sense that regressors and regressands are the only input required from the econometrician, so that they do not need to engage in technically demanding specifications of priors and model parametrizations.
We conducted several Monte Carlo experiments to understand the finite-sample properties of the model. We found that the model had satisfactory estimation precision and forecasting performance even when regression coefficients were stable or when coefficient instability was present but the model was mis-specified. When coefficients were unstable, the estimation precision and the forecasting accuracy of the model were significantly better than those of other competing models.
A caveat emerged from our Monte Carlo experiments: When a highly persistent autoregressive component was included among the regressors, then the TVC model tended to have poor estimation precision; in this case, the performance of the TVC model could be improved by increasing the number of exogenous regressors or by increasing the sample size, otherwise, one could resort to model-averaging variants of frequentist breakpoint detection procedures.
To demonstrate a real-world application of our TVC model, we used it to estimate the exposures of S&P 500 stocks to market-wide risk factors. We found that a vast majority of stocks had time-varying exposures and that the TVC model helped to better forecast these exposures.

Author Contributions

All authors contributed equally to the paper.

Funding

This research received no external funding.

Acknowledgments

We thank Giovanni Urga and Fabrizio Venditti for helpful comments, as well as the audiences at the following seminars/conferences: The Bank of Italy, EIEF, Cass Business School, Scottish economic Society. The views expressed here are our own and do not necessarily reflect those of the Bank of Italy. Other usual disclaimers apply.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Appendix A.1. Proofs of Propositions 1 and 2

In this section we derive the formulae presented in Propositions 1 and 2. To facilitate the exposition, we start from simpler information structures and then we tackle the more complex information structure assumed in Propositions 1 and 2 and summarized in Summary 1.

Appendix A.1.1. V and θ Known, β 1 Unknown

We start from the simple case in which V and θ are both known. The assumptions on the priors and the initial information are summarized as follows:
Case A1
(Priors and initial information). The priors on the unknown parameters are:
β 1 D 0 N β ^ 1 , 0 , V F β , 1 , 0
and the initial information set is:
D 0 = β ^ 1 , 0 , F β , 1 , 0 , V , θ
Note that also W = λ θ F β , 1 , 0 and W = V λ θ F β , 1 , 0 are known, because θ and V are known. The information sets D t satisfy the recursion D t = D t 1 y t , x t , starting from the set D 0 . Given the above assumptions, as new information becomes available, the posterior distribution of the parameters of the regression can be calculated using the following results:
Proposition A1
(Forward updating). Let priors and initial information be as in Case A1. Then:
β t D t 1 N β ^ t , t 1 , V F β , t , t 1 y t D t 1 , x t N y ^ t , t 1 , V F y , t , t 1 β t D t N β ^ t , t , V F β , t , t
where the means and variances of the above distributions are calculated recursively as follows:
β ^ t , t 1 = β ^ t 1 , t 1 F β , t , t 1 = F β , t 1 , t 1 + W y ^ t , t 1 = x t β ^ t , t 1 F y , t , t 1 = 1 + x t F β , t , t 1 x t e t = y t y ^ t , t 1 P t = F β , t , t 1 x t / F y , t , t 1 β ^ t , t = β ^ t , t 1 + P t e t F β , t , t = F β , t , t 1 P t P t F y , t , t 1
starting from the initial values β ^ 1 , 0 and F β , 1 , 0 .
Proof. 
Note that, given the above assumptions, the system:
y t = x t β t + v t β t = β t 1 + w t
is a Gaussian linear state-space system, where y t = x t β t + v t is the observation equation and β t = β t 1 + w t is the transition equation. Hence, the posterior distribution of the states can be updated using the Kalman filter. The recursive equations in Proposition A1 are just the usual updating equations of the Kalman filter (e.g., Hamilton 1994). □
The smoothing equations are provided by the following proposition:
Proposition A2
(Backward updating). Let priors and initial information be as in Case A1. Then:
β T τ D T N β ^ T τ , T , V F β , T τ , T
where the means and the variances of the above distributions are calculated recursively (backwards) as follows:
Q T τ = F β , T τ , T τ F β , T τ + 1 , T τ 1 β ^ T τ , T = β ^ T τ , T τ + Q T τ β ^ T τ + 1 , T β ^ T τ + 1 , T τ F β , T τ , T = F T τ , T τ + Q T τ F T τ + 1 , T F T τ + 1 , T τ Q T τ
and the backward recursions start from the terminal values of the forward recursions calculated in Proposition 5.
Proof. 
These are the usual backward Kalman recursions (e.g., Hamilton 1994). □

Appendix A.1.2. θ Known, β1 and V Unknown

In this subsection we relax the assumption that V (the variance of v t ) is known and we impose a Gamma prior on the reciprocal of V. The assumptions on the priors and the initial information are summarized as follows:
Case A2
(Priors and initial information). The priors on the unknown parameters are:
β 1 D 0 N β ^ 1 , 0 , V F β , 1 , 0 1 / V D 0 G V ^ 0 , n 0
and the initial information set is:
D 0 = β ^ 1 , 0 , F β , 1 , 0 , V ^ 0 , n 0 , θ
Note that also W = λ θ F β , 1 , 0 is known, because θ is known. The information sets D t satisfy the recursion D t = D t 1 y t , x t , starting from the set D 0 . Given the above assumptions, the posterior distributions of the parameters of the regression can be calculated as follows:
Proposition A3
(Forward updating). Let priors and initial information be as in Case A2. Then:
β t D t 1 T β ^ t , t 1 , V ^ t 1 F β , t , t 1 , n t 1 y t D t 1 , x t T y ^ t , t 1 , V ^ t 1 F y , t , t 1 , n t 1 β t D t T β ^ t , t , V ^ t F β , t , t , n t 1 / V D t G V ^ t , n t
where the parameters of the above distributions are calculated recursively as in Proposition A1 and as follows:
n t = n t 1 + 1 V ^ t = 1 n t n t 1 V ^ t 1 + e t 2 F y , t , t 1
starting from the initial values β ^ 1 , 0 , F β , 1 , 0 , V ^ 0 and n 0 .
Proof. 
The proof is by induction. At time t = 1 , p β 1 D 0 , V and p 1 / V D 0 are the conjugate normal/inverse gamma priors of a standard Bayesian regression model with constant coefficients (e.g., Hamilton 1994). Therefore, the usual results on the updating of these conjugate priors hold:
β 1 D 1 , V N β ^ 1 , 1 , V F β , 1 , 1
1 / V D 1 G V ^ 1 , n 1
Since β 2 = β 1 + w 2 and
w 2 D 1 , V N 0 , V W
then, by the additivity of normal distributions:
β 2 D 1 , V N β ^ 1 , 1 , V F β , 1 , 1 + V W = N β ^ 2 , 1 , V F β , 2 , 1
Therefore, at time t = 2 , p β 2 D 1 , V and p 1 / V D 1 are again the conjugate normal/inverse gamma priors of a standard Bayesian regression model with constant coefficients. Proceeding in the same way as for t = 1 , one obtains the desired result for t = 2 and, inductively, for all the other periods. □
Posterior distributions of the coefficients that take into account all information received up to time T are calculated as follows:
Proposition A4
(Backward updating). Let priors and initial information be as in Case A2. Then:
β T τ D T T β ^ T τ , T , V ^ T F β , T τ , T , n T
where V ^ T and n T are calculated as in Proposition A3 and the other parameters of the above distributions are calculated recursively (backwards) as in Proposition A2.
Proof. 
From Proposition A2, we know that:
β T τ D T , V = β T τ D T N β ^ T τ , T , V F β , T τ , T
Furthermore, 1 / V D T G V ^ T , n T . By the conjugacy of β T τ D T , V and 1 / V D T , it follows that:
β T τ D T T β ^ T τ , T , V ^ T F β , T τ , T , n T
 □

Appendix A.1.3. θ, β1 and V Unknown

In this subsection we relax the assumption that θ is known, using the same priors and initial information of the propositions in the main text of the article (Propositions 1 and 2):
Case A3.
The priors on the unknown parameters are:
β 1 D 0 , V , θ N β ^ 1 , 0 , V F β , 1 , 0 1 / V D 0 , θ G V ^ 0 , n 0 p θ i D 0 = p 0 , i , i = 1 , , q
and the initial information set is:
D 0 = β ^ 1 , 0 , F β , 1 , 0 , V ^ 0 , n 0 , p 0 , 1 , , p 0 q
The information sets D t satisfy the recursion D t = D t 1 y t , x t , starting from the set D 0 . Note that the assumptions introduced in Cases A1 and A2 in the previous subsections had the sole purpose of introducing the more complex Case A3. Given the above assumptions, the posterior distributions of the parameters of the regression can be calculated as follows:
Proposition A5.
Let priors and initial information be as in Case A3. Let p t , i = p θ = θ i D t . Then:
p β t D t 1 = i = 1 q p β t θ = θ i , D t 1 p t 1 , i p y t D t 1 , x t = i = 1 q p y t θ = θ i , D t 1 , x t p t 1 , i p 1 / V D t 1 = i = 1 q p 1 / V θ = θ i , D t 1 p t 1 , i p β t D t = i = 1 q p β t θ = θ i , D t p t , i
The mixing probabilities are obtained recursively as:
p t , i = p t 1 , i p y t θ = θ i , D t 1 , x t j = 1 q p t 1 , j p y t θ = θ j , D t 1 , x t
starting from the prior probabilities p 0 , 1 , , p 0 , q . The conditional densities:
p β t D t 1 , θ = θ i = p β t D t 1 p y t D t 1 , x t , θ = θ i = p y t D t 1 , x t p 1 / V D t 1 , θ = θ i = p 1 / V D t 1 p β t D t , θ = θ i = p β t D t
are calculated for each θ i as in Propositions A1 and A3.
Proof. 
Conditioning on θ = θ i , the distributions of the parameters β t and V and of the observations y t are obtained from Propositions A1 and A3 (it suffices to note that D t θ = D t ). Not conditioning on θ = θ i , the distributions of the parameters β t and V and of the observations y t are obtained marginalizing their joint distribution with θ . For example:
p β t D t 1 = i = 1 q p β t , θ i D t 1 = i = 1 q p β t θ = θ i , D t 1 p θ = θ i D t 1 = i = 1 q p β t θ = θ i , D t 1 p t 1 , i
The mixing probabilities are obtained using Bayes’ rule:
p t , i = p θ = θ i D t = p θ = θ i y t , D t 1 , x t = p y t θ = θ i , D t 1 , x t p θ = θ i D t 1 , x t p y t D t 1 , x t = p y t θ = θ i , D t 1 , x t p θ = θ i D t 1 j = 1 q p y t , θ j D t 1 , x t = p y t θ = θ i , D t 1 , x t p θ = θ i D t 1 j = 1 q p y t θ = θ j , D t 1 , x t p θ = θ j D t 1 , x t = p y t θ = θ i , D t 1 , x t p t 1 , i j = 1 q p y t θ = θ j , D t 1 , x t p t 1 , j
 □
Proposition 1 in the main text is obtained by combining Propositions A1, A3, and A5 above. Proposition 2 results from Propositions A2, A4, and A5 above.

References

  1. Andrews, Donald W. K. 1993. Tests for parameter instability and structural change with unknown change point. Econometrica 61: 821–56. [Google Scholar] [CrossRef]
  2. Andrews, Donald W. K., Inpyo Lee, and Werner Ploberger. 1996. Optimal changepoint tests for normal linear regression model. Journal of Econometrics 70: 9–38. [Google Scholar] [CrossRef]
  3. Bai, Jushan, and Pierre Perron. 1998. Estimating and testing linear models with multiple structural breaks. Econometrica 66: 47–78. [Google Scholar] [CrossRef]
  4. Bai, Jushan, and Pierre Perron. 2003. Computation and analysis of multiple structural change models. Journal of Applied Econometrics 18: 1–22. [Google Scholar] [CrossRef]
  5. Bai, Jushan, and Pierre Perron. 2006. Multiple structural change models: A simulation analysis. In Econometric Theory and Practice: Frontier of Analysis and Applied Research (Essays in Honor of Peter Phillips). Edited by D. Corbae, S. Durlauf and B. E. Hansen. Cambridge: Cambridge University Press. [Google Scholar]
  6. Bitto, Angela, and Sylvia Frühwirth-Schnatter. 2019. Achieving shrinkage in a time-varying parameter model framework. Journal of Econometrics 210: 75–97. [Google Scholar] [CrossRef]
  7. Brown, Robert L., James Durbin, and James M. Evans. 1975. Techniques for testing the constancy of regression coefficients over time. Journal of the Royal Statistical Society B 37: 149–92. [Google Scholar]
  8. Carter, Chris K., and Robert Kohn. 1994. On Gibbs sampling for state space models. Biometrika 81: 541–53. [Google Scholar] [CrossRef]
  9. Casarin, Roberto, German Molina, and Enrique ter Horst. 2019. A Bayesian time varying approach to risk neutral density estimation. Journal of the Royal Statistical Society: Series A 182: 165–95. [Google Scholar] [CrossRef]
  10. Chan, Joshua C. C. 2017. The stochastic volatility in mean model with time-varying parameters: An application to inflation modeling. Journal of Business & Economic Statistics 35: 17–28. [Google Scholar]
  11. Chib, Siddhartha, and Edward Greenberg. 1995. Hierarchical analysis of SUR models with extensions to correlated serial errors and time-varying parameter models. Journal of Econometrics 68: 339–60. [Google Scholar] [CrossRef]
  12. Chow, Gregory C. 1960. Tests of equality between sets of coefficients in two linear regressions. Econometrica 28: 591–605. [Google Scholar] [CrossRef]
  13. Cogley, Timothy, Giorgio E. Primiceri, and Thomas J. Sargent. 2010. Inflation-gap persistence in the US. American Economic Journal: Macroeconomics 2: 43–69. [Google Scholar] [CrossRef]
  14. Cogley, Timothy, and Thomas J. Sargent. 2001. Evolving Post World War II U.S. Inflation Dynamics. NBER Macroeconomics Annual 16: 331–73. [Google Scholar]
  15. Datta, Gauri Sankar. 1996. On priors providing frequentist validity of Bayesian inference for multiple parametric functions. Biometrika 83: 287–98. [Google Scholar] [CrossRef]
  16. Datta, Gauri Sankar, and Jayanta Kumar Ghosh. 1995. On priors providing frequentist validity for Bayesian inference. Biometrika 82: 37–45. [Google Scholar] [CrossRef]
  17. D’Agostino, Antonello, Luca Gambetti, and Domenico Giannone. 2013. Macroeconomic forecasting and structural change. Journal of Applied Econometrics 28: 82–101. [Google Scholar] [CrossRef]
  18. Doan, Thomas, Robert Litterman, and Christopher Sims. 1984. Forecasting and conditional projection using realistic prior distributions. Econometric Reviews 3: 1–100. [Google Scholar] [CrossRef] [Green Version]
  19. Fama, Eugene F., and Kenneth R. French. 1993. Common risk factors in the returns on bonds and stocks. Journal of Financial Economics 33: 3–56. [Google Scholar] [CrossRef]
  20. Fama, Eugene F., and Kenneth R. French. 1995. Size and book-to-market factors in earnings and returns. Journal of Finance 50: 131–55. [Google Scholar] [CrossRef]
  21. Fama, Eugene F., and Kenneth R. French. 1996. Multifactor explanations of asset pricing anomalies. Journal of Finance 51: 55–84. [Google Scholar] [CrossRef]
  22. Fernandez, Carmen, Eduardo Ley, and Mark F. J. Steel. 2001. Benchmark priors for Bayesian model averaging. Journal of Econometrics 100: 381–427. [Google Scholar] [CrossRef] [Green Version]
  23. George, Edward I., and Robert E. McCulloch. 1997. Approaches for Bayesian variable selection. Statistica Sinica 7: 339–74. [Google Scholar]
  24. Ghosh, J. K., and Rahul Mukerjee. 1993. Frequentist validity of highest posterior density regions in the multiparameter case. Annals of the Institute of Statistical Mathematics 45: 293–302. [Google Scholar] [CrossRef]
  25. Guerre, Emmanuel, and Pascal Lavergne. 2005. Data-driven rate-optimal specification testing in regression models. The Annals of Statistics 33: 840–70. [Google Scholar] [CrossRef]
  26. Hamilton, James D. 1994. Time Series Analysis. Princeton: Princeton University Press. [Google Scholar]
  27. Hansen, Bruce E. 2000. Testing for structural change in conditional models. Journal of Econometrics 97: 93–115. [Google Scholar] [CrossRef] [Green Version]
  28. Hatanaka, Michio, and Kazuo Yamada. 1999. A unit root test in the presence of structural changes in I(1) and I(0) models. In Cointegration, Causality and Forecasting: A Festschrift in Honour of Clive W. J Granger. Edited by R. F. Engle and H. White. Oxford: Oxford University Press. [Google Scholar]
  29. Hoeting, Jennifer A., David Madigan, Adrian E. Raftery, and Chris T. Volinsky. 1999. Bayesian model averaging: a tutorial. Statistical Science 14: 382–401. [Google Scholar]
  30. Horowitz, Joel L., and Vladimir G. Spokoiny. 2001. An adaptive, rate-optimal test of a parametric model against a nonparametric alternative. Econometrica 69: 599–631. [Google Scholar] [CrossRef]
  31. Jeffreys, Harold. 1961. The Theory of Probability, 3rd ed. Oxford: Oxford University Press. [Google Scholar]
  32. Kass, Robert E., and Larry Wasserman. 1995. A reference Bayesian test for nested hypotheses and its relationship to the Schwarz criterion. Journal of the American Statistical Association 90: 928–34. [Google Scholar] [CrossRef]
  33. Kastner, Gregor. 2019. Sparse Bayesian time-varying covariance estimation in many dimensions. Journal of Econometrics 210: 98–115. [Google Scholar] [CrossRef]
  34. Lepski, Oleg V., Enno Mammen, and Vladimir G. Spokoiny. 1997. Optimal spatial adaptation to inhomogeneous smoothness: An approach based on kernel estimates with variable bandwidth selectors. Annals of Statistics 25: 929–47. [Google Scholar]
  35. Liang, Feng, Rui Paulo, German Molina, Merlise A. Clyde, and Jim O. Berger. 2008. Mixtures of g priors for Bayesian variable selection. Journal of the American Statistical Association 103: 410–23. [Google Scholar] [CrossRef]
  36. Mukerjee, Rahul, and Dipak K. Dey. 1993. Frequentist validity of posterior quantiles in the presence of a nuisance parameter: Higher order asymptotics. Biometrika 80: 499–505. [Google Scholar] [CrossRef]
  37. Mumtaz, Haroon, and Konstantinos Theodoridis. 2018. The changing transmission of uncertainty shocks in the US. Journal of Business & Economic Statistics 36: 239–52. [Google Scholar]
  38. Nyblom, Jukka. 1989. Testing for the constancy of parameters over time. Journal of the American Statistical Association 84: 348–68. [Google Scholar] [CrossRef]
  39. Pacifico, Antonio. 2019. Structural Panel Bayesian VAR Model to Deal with Model Misspecification and Unobserved Heterogeneity Problems. Econometrics 7: 8. [Google Scholar] [CrossRef]
  40. Perron, Pierre, and Xiaokang Zhu. 2005. Structural breaks with deterministic and stochastic trends. Journal of Econometrics 129: 65–119. [Google Scholar] [CrossRef]
  41. Pesaran, M. Hashem, and Allan Timmermann. 2002. Market timing and return prediction under model instability. Journal of Empirical Finance 9: 495–510. [Google Scholar] [CrossRef] [Green Version]
  42. Pesaran, M. Hashem, and Allan Timmermann. 2007. Selection of estimation window in the presence of breaks. Journal of Econometrics 137: 134–61. [Google Scholar] [CrossRef]
  43. Petrova, Katerina. 2019. A quasi-Bayesian local likelihood approach to time varying parameter VAR models. Journal of Econometrics. [Google Scholar] [CrossRef]
  44. Raftery, Adrian E., Miroslav Kárný, and Pavel Ettler. 2010. Online prediction under model uncertainty via dynamic model averaging: Application to a cold rolling mill. Technometrics 52: 52–66. [Google Scholar] [CrossRef]
  45. Robert, Christian. 2007. The Bayesian Choice: From Decision-Theoretic Foundations to Computational Implementation. Berlin: Springer. [Google Scholar]
  46. Shively, Thomas S., Robert Kohn, and Sally Wood. 1999. Variable selection and function estimation in additive nonparametric regression using a data-based prior (with discussion). Journal of the American Statistical Association 94: 777–806. [Google Scholar] [CrossRef]
  47. Smith, Michael, and Robert Kohn. 1996. Nonparametric regression using Bayesian variable selection. Journal of Econometrics 75: 317–43. [Google Scholar] [CrossRef]
  48. Spokoiny, Vladimir. 2001. Data-driven testing the fit of linear models. Mathematical Methods of Statistics 10: 465–97. [Google Scholar]
  49. Stock, James H., and Mark W. Watson. 1996. Evidence on structural instability in macroeconomic time series relations. Journal of Business & Economic Statistics 14: 11–30. [Google Scholar]
  50. West, Mike, and Jeff Harrison. 1997. Bayesian Forecasting and Dynamic Models, 2nd ed. New York: Springer. [Google Scholar]
  51. Zellner, Arnold. 1986. On assessing prior distributions and Bayesian regression analysis with g-prior distributions. In Bayesian Inference and Decision Techniques: Essays in Honor of Bruno de Finetti. Edited by P. K. Goel and A. Zellner. Amsterdam: North-Holland and Elsevier, pp. 233–43. [Google Scholar]
1.
An alternative approach is to specify and estimate regression models under the hypothesis of constant coefficients, and then test for the presence of structural breaks (e.g., Chow 1960; Brown et al. 1975; Nyblom 1989) and identify the breakpoints (e.g., Andrews 1993; Andrews et al. 1996; Bai and Perron 1998).
2.
For Monte Carlo studies of frequentist breakpoint detection methods see Hansen (2000) and Bai and Perron (2006).
3.
Before arriving to the specification of priors proposed in this paper, we tried several other specifications and found that the results can indeed be quite sensitive to rescalings if one chooses other priors.
4.
A MATLAB function is made available on the internet at https://www.statlect.com/time_varying_regression.htm. The function can be called with the instruction:
tvc ( y , X )
where y is a T × 1 vector of observations on the dependent variable and X is a T × K matrix of regressors.
5.
However, in their model the coefficients β t do not follow a random walk (they are mean reverting). They also use different priors: While we impose Zellner’s g-prior on β 1 (see Section 2), they impose the Minnesota prior.
6.
However, they assume that F β , 1 , 0 is proportional to the identity matrix, while we assume that also F β , 1 , 0 is proportional to X X 1 . Furthermore, they do not estimate V. Their analysis is focused on the one-step-ahead predictions of y t , which can be computed without knowing V. They approach the estimation of λ in a number of different ways, but none of them allows one to derive analytically a posterior distribution for λ .
7.
In their model the prior covariance of w t is proportional to X X 1 , but X is the design matrix of a pre-sample not used for the estimation of the model.
8.
For example, in the empirical part of the paper, setting q = 100 and c = 0.9 , we are able to simultaneously consider 5 different orders of magnitude of instability. With the same number of points q and an arithmetic grid, we would have been able to consider only 2 orders.
9.
Note, however, that the parallelism can be misleading, as Bayesian p-values have only a frequentist validity in special cases. Ghosh and Mukerjee (1993); Mukerjee and Dey (1993); Datta and Ghosh (1995), and Datta (1996) provide conditions that priors have to satisfy in order for Bayesian p-values to have also frequentist validity.
10.
We estimate the breakpoint dates sequentially rather than simultaneously to achieve a reasonable computational speed in our Monte Carlo simulations. Denote by ν S the number of breakpoints estimated by the sequential procedure and by ν σ the number estimated by the simultaneous procedure. Given that we are using the SIC criterion to choose the number of points, if ν σ 1 , then ν S = ν σ ; otherwise, if ν σ > 1 , then ν S ν σ . Therefore, in our Monte Carlo simulations (where the true number of breakpoints is either 0 or 1), the sequential procedure provides a better estimate of the number of breakpoints than the simultaneous procedure.
11.
This identification problem is discussed in a very similar context by Hatanaka and Yamada (1999) and Perron and Zhu (2005).
Table 1. Estimation precision when coefficients are stable—Monte Carlo evidence—MSE of coefficient estimates.
Table 1. Estimation precision when coefficients are stable—Monte Carlo evidence—MSE of coefficient estimates.
Panel A—One Lag in the Estimated Equation
TVC-MATVC-MSTVC-PTVC-pOLSBPBP-MA
ρ = 0
T = 1000.02570.02860.02390.02150.02050.05400.0221
T = 2000.01380.01420.01210.01060.01020.04790.0108
T = 5000.00660.00560.00490.00430.00400.00700.0043
ρ = 0.50
T = 1000.02710.03030.02490.02190.02070.07300.0227
T = 2000.01360.01410.01190.01040.00990.03290.0106
T = 5000.00640.00550.00480.00410.00380.01320.0039
ρ = 0.80
T = 1000.03080.03520.02780.02340.02190.15370.0272
T = 2000.01410.01450.01210.01030.00970.08530.0105
T = 5000.00620.00530.00460.00390.00370.01790.0038
ρ = 0.99
T = 1004.79335.34984.72334.73440.06661.92220.1145
T = 2002.53902.72972.50722.50570.02440.71890.0367
T = 5000.44480.47920.43010.42870.00620.13380.0072
Panel B—Three Lags in the Estimated Equation
TVC-MATVC-MSTVC-PTVC-pOLSBPBP-MA
ρ = 0
T = 1000.08850.09640.08600.08040.07850.08270.0795
T = 2000.04330.04530.04140.03850.03760.03790.0378
T = 5000.01860.01750.01620.01500.01460.01460.0146
ρ = 0.50
T = 1000.09530.10330.09200.08630.08420.09050.0850
T = 2000.04570.04780.04340.04060.03980.03990.0398
T = 5000.01940.01840.01700.01590.01550.01570.0156
ρ = 0.80
T = 1000.10490.11340.10170.09590.09400.10240.0947
T = 2000.05110.05320.04840.04540.04420.04470.0443
T = 5000.02150.02030.01880.01740.01690.01690.0169
ρ = 0.99
T = 1000.72130.95990.62440.67130.15050.18300.1531
T = 2000.26190.30520.23170.24250.06320.08670.0636
T = 5000.03270.03100.02790.02640.02130.02130.0213
Table 2. Prediction accuracy when coefficients are stable—Monte Carlo evidence—MSE of one-step-ahead predictions.
Table 2. Prediction accuracy when coefficients are stable—Monte Carlo evidence—MSE of one-step-ahead predictions.
Panel A—One Lag in the Estimated Equation
TVC-MATVC-MSTVC-PTVC-pOLSBPBP-MA
ρ = 0
T = 1001.03841.04381.03581.03191.03011.09421.0332
T = 2001.02081.02141.01811.01611.01551.29141.0165
T = 5001.01011.00841.00731.00641.00601.01211.0073
ρ = 0.50
T = 1001.03931.04431.03661.03251.03111.09781.0351
T = 2001.02171.02291.01931.01701.01591.04431.0168
T = 5001.01011.00861.00751.00651.00601.01441.0061
ρ = 0.80
T = 1001.04531.05261.04231.03711.03481.13011.0420
T = 2001.02181.02301.01911.01621.01551.07571.0166
T = 5001.01031.00891.00781.00651.00621.02291.0063
ρ = 0.99
T = 1001.22561.27101.22381.22061.04891.21991.0484
T = 2001.13491.15481.13171.13061.02241.04961.0225
T = 5001.04191.04621.03921.03771.00781.01981.0078
Panel B—Three Lags in the Estimated Equation
TVC-MATVC-MSTVC-PTVC-pOLSBPBP-MA
ρ = 0
T = 1001.08411.09141.08131.07621.07461.08081.0759
T = 2001.04331.04541.04121.03861.03751.03791.0379
T = 5001.01831.01711.01581.01461.01421.01431.0143
ρ = 0.50
T = 1001.08451.09231.08171.07691.07521.09281.0772
T = 2001.04351.04541.04121.03801.03721.03751.0375
T = 5001.01791.01671.01541.01441.01411.01411.0141
ρ = 0.80
T = 1001.08891.09671.08641.08161.08021.08551.0815
T = 2001.04181.04351.03971.03731.03631.03691.0367
T = 5001.01771.01681.01551.01431.01391.01401.0140
ρ = 0.99
T = 1001.13811.17051.12981.13531.09431.11281.0886
T = 2001.06471.07271.06091.06171.04391.05041.0425
T = 5001.01961.01921.01781.01711.01631.01601.0160
Table 3. Estimation precision when coefficients experience one discrete break—Monte Carlo evidence—MSE of coefficient estimates.
Table 3. Estimation precision when coefficients experience one discrete break—Monte Carlo evidence—MSE of coefficient estimates.
Panel A—One Lag in the Estimated Equation
TVC-MATVC-MSTVC-PTVC-pOLSBPBP-MA
ρ = 0
T = 1000.17720.17670.18070.19430.37690.17370.2252
T = 2000.11480.11450.11830.12500.34340.10880.1863
T = 5000.06890.06920.07110.07410.34530.04520.1738
ρ = 0.50
T = 1000.17900.17920.18310.19510.36550.21400.2248
T = 2000.12370.12280.12650.13460.35450.10530.1930
T = 5000.06920.06940.07130.07440.34520.05330.1733
ρ = 0.80
T = 1000.18870.19060.19240.20380.37050.30160.2226
T = 2000.12610.12550.12900.13600.34630.13090.1894
T = 5000.07070.07080.07290.07600.34490.08030.1733
ρ = 0.99
T = 1005.18685.77415.15085.16180.42832.52680.4261
T = 2003.30693.50543.29083.29560.36261.46810.2585
T = 5000.70540.73620.70320.70530.35420.64310.1946
Panel B—Three Lags in the Estimated Equation
TVC-MATVC-MSTVC-PTVC-pOLSBPBP-MA
ρ = 0
T = 1000.32990.33680.33220.34240.43890.36280.3320
T = 2000.21270.21440.21450.22020.37610.21860.2421
T = 5000.13570.13610.13720.14000.35400.10530.1974
ρ = 0.50
T = 1000.32720.33470.32910.33770.43680.39480.3292
T = 2000.22010.22180.22150.22700.38260.21940.2460
T = 5000.14010.14050.14130.14350.35070.10780.1956
ρ = 0.80
T = 1000.35290.36270.35470.36280.45850.41610.3476
T = 2000.24200.24410.24360.24990.39130.25010.2591
T = 5000.14750.14780.14850.15130.36380.11550.2040
ρ = 0.99
T = 1001.36421.62161.31551.34680.53191.38140.5141
T = 2000.73990.80440.72790.73890.42960.76780.3266
T = 5000.25320.25690.25410.25650.36510.31680.2147
Table 4. Prediction accuracy when coefficients experience one discrete break—Monte Carlo evidence—MSE of one-step-ahead predictions.
Table 4. Prediction accuracy when coefficients experience one discrete break—Monte Carlo evidence—MSE of one-step-ahead predictions.
Panel A—One Lag in the Estimated Equation
TVC-MATVC-MSTVC-PTVC-pOLSBPBP-MA
ρ = 0
T = 1001.29751.29891.30301.32431.60291.33221.3784
T = 2001.18101.17941.18671.19871.53151.17771.2937
T = 5001.11011.11051.11381.12081.61741.09301.3093
ρ = 0.50
T = 1001.26451.26431.27531.29911.54211.29411.3343
T = 2001.20311.20391.20951.21921.64921.17001.3604
T = 5001.10991.11061.11411.12131.61621.09081.3064
ρ = 0.80
T = 1001.28141.28151.28821.32021.57811.38591.3552
T = 2001.16931.16841.17421.18761.56741.14811.2955
T = 5001.10821.10851.11261.11871.61581.08671.3069
ρ = 0.99
T = 1001.40141.43571.40511.42451.64421.39331.3764
T = 2001.28211.29511.28411.29291.54191.20051.3027
T = 5001.18071.18571.18461.18871.68441.15521.3481
Panel B—Three Lags in the Estimated Equation
TVC-MATVC-MSTVC-PTVC-pOLSBPBP-MA
ρ = 0
T = 1001.40071.40221.40421.42721.61171.54081.4496
T = 2001.26891.26921.27381.28551.57371.33081.3612
T = 5001.17701.17731.18031.18771.57531.11501.3172
ρ = 0.50
T = 1001.40221.40591.40701.42551.64901.65501.4823
T = 2001.28321.28351.28721.29691.60131.35151.3825
T = 5001.15761.15821.16001.16341.54361.11251.2884
ρ = 0.80
T = 1001.42061.42451.42421.44261.69301.51601.4954
T = 2001.28641.28311.28911.30451.60341.27471.3675
T = 5001.17331.17271.17581.18201.63481.11531.3417
ρ = 0.99
T = 1001.47181.50541.48521.51341.66441.45211.4802
T = 2001.31231.32241.31831.33651.61581.36701.3750
T = 5001.19161.19341.19671.20291.59661.11381.3327
Table 5. Estimation precision when coefficients change every period—Monte Carlo evidence—MSE of coefficient estimates.
Table 5. Estimation precision when coefficients change every period—Monte Carlo evidence—MSE of coefficient estimates.
Panel A—One Lag in the Estimated Equation
TVC-MATVC-MSTVC-PTVC-pOLSBPBP-MA
ρ = 0
T = 1000.17680.17780.18030.19390.36530.23330.2331
T = 2000.12070.12100.12240.12800.35350.19140.2048
T = 5000.07180.07180.07220.07350.34090.09250.1766
ρ = 0.50
T = 1000.18560.18650.18910.20310.37750.26890.2413
T = 2000.12160.12190.12340.12870.35310.16350.2055
T = 5000.07190.07200.07240.07350.34900.09820.1770
ρ = 0.80
T = 1000.20030.20420.20290.21460.37380.32520.2403
T = 2000.12680.12750.12850.13330.33530.19820.1971
T = 5000.07330.07330.07370.07470.34130.11080.1759
ρ = 0.99
T = 1005.88726.59305.86555.86740.44813.80890.4898
T = 2003.24783.41243.24353.24280.37661.63240.2718
T = 5000.86090.89160.86080.86090.33950.58050.1835
TVC-MATVC-MSTVC-PTVC-pOLSBPBP-MA
ρ = 0
T = 1000.32790.33780.33040.34070.43860.39030.3544
T = 2000.21870.22210.22020.22470.38100.25140.2723
T = 5000.13490.13580.13540.13640.35260.15590.2151
ρ = 0.50
T = 1000.33690.34830.33910.34780.43950.39540.3553
T = 2000.22790.23130.22920.23390.38340.26070.2753
T = 5000.13890.13970.13930.14040.35530.16080.2176
ρ = 0.80
T = 1000.36080.37350.36320.37280.45000.45280.3665
T = 2000.23970.24420.24100.24560.39010.27560.2806
T = 5000.14890.14980.14930.15040.36550.16860.2241
ρ = 0.99
T = 1001.36791.65641.31191.34570.53451.21360.5107
T = 2000.77020.83730.76480.76900.42310.55110.3365
T = 5000.24670.25110.24720.24800.36330.23230.2291
Table 6. Prediction accuracy when coefficients change every period—Monte Carlo evidence—MSE of one-step-ahead predictions.
Table 6. Prediction accuracy when coefficients change every period—Monte Carlo evidence—MSE of one-step-ahead predictions.
Panel A—One Lag in the Estimated Equation
TVC-MATVC-MSTVC-PTVC-pOLSBPBP-MA
ρ = 0
T = 1001.26371.26601.26961.28871.57931.36561.3679
T = 2001.18251.18331.18581.19421.56271.26221.3261
T = 5001.10941.10961.11011.11191.55461.15151.2784
ρ = 0.50
T = 1001.27601.27511.28301.30931.59481.41551.3831
T = 2001.18671.18681.18941.20001.55541.24331.3298
T = 5001.11611.11591.11701.11921.59431.15641.3054
ρ = 0.80
T = 1001.30091.30261.30701.33391.61251.42571.4093
T = 2001.18941.19031.19261.20471.54931.27231.3286
T = 5001.11161.11151.11231.11391.53991.16721.2843
ρ = 0.99
T = 1001.40281.43631.40671.42201.61461.49921.3924
T = 2001.28481.29721.28671.29171.59631.26751.3379
T = 5001.15641.15961.15741.15881.55811.17521.2882
TVC-MATVC-MSTVC-PTVC-pOLSBPBP-MA
ρ = 0
T = 1001.43121.43331.43591.46361.66671.50281.5340
T = 2001.26531.26531.26811.28001.55771.31861.3912
T = 5001.17131.17151.17231.17371.57451.21561.3453
ρ = 0.50
T = 1001.42141.42471.42521.44761.63731.49821.5109
T = 2001.29541.29611.29931.31141.62801.36971.4369
T = 5001.16871.16891.16951.17111.56311.21451.3402
ρ = 0.80
T = 1001.43031.43421.43531.46141.66161.55341.5281
T = 2001.28031.28161.28291.29411.59491.33011.4126
T = 5001.16521.16481.16591.16811.55841.22401.3402
ρ = 0.99
T = 1001.51381.54171.52461.56641.68481.55741.5380
T = 2001.34051.35331.34691.36241.62281.35071.4279
T = 5001.18381.18511.18531.18891.58421.24451.3475
Table 7. Prediction accuracy—Risk exposures—Reduction in the MSE of one-step-ahead predictions (benchmark = OLS).
Table 7. Prediction accuracy—Risk exposures—Reduction in the MSE of one-step-ahead predictions (benchmark = OLS).
TVC-MATVC-MSTVC-PTVC-pOLSBPBP-MA
Mean2.94%2.48%2.70%2.15%0%−121.72%3.13%
Standard dev.11.02%11.68%10.82%10.63%0%−557.07%6.00%
First quartile−2.14%−2.67%−2.14%−2.23%0%−46.94%−0.14%
Median1.85%1.30%1.17%0.11%0%−5.76%1.53%
Third quartile7.83%7.75%7.47%6.84%0%0.20%6.37%

Share and Cite

MDPI and ACS Style

Ciapanna, E.; Taboga, M. Bayesian Analysis of Coefficient Instability in Dynamic Regressions. Econometrics 2019, 7, 29. https://0-doi-org.brum.beds.ac.uk/10.3390/econometrics7030029

AMA Style

Ciapanna E, Taboga M. Bayesian Analysis of Coefficient Instability in Dynamic Regressions. Econometrics. 2019; 7(3):29. https://0-doi-org.brum.beds.ac.uk/10.3390/econometrics7030029

Chicago/Turabian Style

Ciapanna, Emanuela, and Marco Taboga. 2019. "Bayesian Analysis of Coefficient Instability in Dynamic Regressions" Econometrics 7, no. 3: 29. https://0-doi-org.brum.beds.ac.uk/10.3390/econometrics7030029

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop