Next Article in Journal / Special Issue
On the Forecast Combination Puzzle
Previous Article in Journal
Consequences of Model Misspecification for Maximum Likelihood Estimation with Missing Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Combination Method for Averaging OLS and GLS Estimators †

1
Department of Economics, Otaru University of Commerce, Otaru 047-8501, Japan
2
Discipline of Business Analytics, The University of Sydney Business School, The University of Sydney, Sydney, NSW 2006, Australia
*
Author to whom correspondence should be addressed.
The authors thank Okui Ryo, Mototsugu Shintani, Arihiro Yoshimura and the conference participants at the 2014 European Meeting of the Econometric Society and the 2013 Kansai Keiryo Keizaigaku Kenkyukai for their helpful comments. Liu acknowledges financial support from the JSPS Grant-in-Aid for Young Scientists (B) No. 25780148 and JSPS KAKENHI Grant (C) No. JP16K03590 and No. JP19K01582. The authors are grateful to the Editor and two reviewers for their constructive comments.
Submission received: 6 June 2019 / Revised: 3 September 2019 / Accepted: 4 September 2019 / Published: 9 September 2019
(This article belongs to the Special Issue Bayesian and Frequentist Model Averaging)

Abstract

:
To avoid the risk of misspecification between homoscedastic and heteroscedastic models, we propose a combination method based on ordinary least-squares (OLS) and generalized least-squares (GLS) model-averaging estimators. To select optimal weights for the combination, we suggest two information criteria and propose feasible versions that work even when the variance-covariance matrix is unknown. The optimality of the method is proven under some regularity conditions. The results of a Monte Carlo simulation demonstrate that the method is adaptive in the sense that it achieves almost the same estimation accuracy as if the homoscedasticity or heteroscedasticity of the error term were known.

1. Introduction

Model averaging has been developed as an alternative to model selection. In many situations, model-averaging methods perform better than alternative model-selection methods. The main reason for this is that model selection delivers a pretest estimator that has inferior properties, and its use can be harmful (see Danilov and Magnus, 2004). Yuan and Yang (2005) provided a detailed discussion on the choice between model averaging and model selection. As one of the pioneers of frequentist model averaging, Hansen (2007) proposed Mallows model averaging (MMA) based on the ordinary least-squares (OLS) estimator for linear regression models with homoscedastic errors. Wan et al. (2010) extended the results for non-nested models with homoscedastic errors. Zhao et al. (2018) is the most recent work in this area. For linear regression models with heteroscedastic errors, Hansen and Racine (2012), Liu and Okui (2013) and Zhang et al. (2013, 2015) proposed model averaging methods that are still based on the OLS estimator, while Liu et al. (2016) proposed a method based on the generalized least squares estimator (GLS). They demonstrated that their methods are optimal in the sense of Li (1987) for homoscedastic or heteroscedastic models. For model averaging in big datasets, Xie (2017) proposed the use of model screening (before averaging) in order to deal with the large number of candidate models/regressors.
However, all previous papers assumed that it was known whether the errors of the true data-generating process are homoscedastic or heteroscedastic. Due to this assumption, the previous averaging methods were based only on estimators constructed using the same estimation method, either OLS or GLS estimators (with different regressor sets). This assumption can be unrealistic in empirical applications. Usually, researchers do not know the structure of the error term; therefore, this assumption leads to possible misspecification. A natural solution is to combine OLS and GLS estimators. Combinations of different methods are routinely used in the applied forecast combination literature. In a recent forecasting competition that included 100,000 series, Makridakis et al. (2018) found that, out of the 17 most accurate methods, 12 were combinations. All combinations used different models/methods varying from simple exponential smoothing models to sophisticated machine-learning algorithms.
We propose a combination method based on OLS and GLS estimators to reduce the risk of misspecification between homoscedastic and heteroscedastic linear models. More precisely, the proposed estimator is a weighted average of mixtures of OLS and GLS estimators. The OLS mixture is constructed using the MMA of Hansen (2007) or the heteroscedasticity-robust Cp(HRCp) model averaging of Liu and Okui (2013). The GLS mixture is constructed using the GLS model averaging (GLSMA) of Liu et al. (2016).
We propose the use of two criteria, MMA-GLSMA and HRCp-GLSMA, to choose the weight vector for combining estimators. The optimality of the chosen weight vector in the sense of Li (1987) is investigated. Our method works in situations with an unknown variance-covariance matrix of the error term if an estimate based on the nonparametric method k-nearest neighbours (k-NN) is used. The results of the simulation experiments show that our combination method is adaptive in the sense that it can achieve almost the same estimation accuracy as if the homoscedasticity or heteroscedasticity of the error term was known.
The rest of the paper is organized as follows: In Section 2, we describe the theoretical setup and introduce the new combination method with the criteria for choosing the weight vector. In Section 3, we investigate the optimality of the proposed criteria. Section 4 presents the results of the Monte Carlo simulations. Section 5 concludes the paper, and all proofs are provided in the Appendix A.

2. Method

Suppose that we have an independent random sample of ( y i , x i ) for i = 1 , , n , where x i = ( x i 1 , x i 2 , , ) is a countably infinite real-valued vector and y i is a real-valued scalar random variable generated from an infinite dimensional linear regression model:
y i = μ i + e i ,
where:
μ i = j = 1 θ j x i j ,
e i is an unobserved error term that can be homoscedastic or heteroscedastic with E ( e i | x i ) = 0 , E ( e i 2 ) = σ 2 and E ( e i 2 | x i ) = σ i 2 and θ j for j = 1 , 2 , are unknown parameters. We also define X x 1 , , x n , Y y 1 , , y n , μ μ 1 , , μ n , e e 1 , , e n and denote the variance-covariance matrix as Ω E e e | X = d i a g ( σ 1 2 , , σ n 2 ) . We state the theoretical results considering the distributions conditional on X and omit all notations for those conditional on X hereafter.

2.1. Infeasible Combination Estimator and Information Criteria

Suppose Ω is known; we have a candidate set of M 1 linear models with different numbers of independent variables for OLS estimation and a candidate set of M 2 linear models with different numbers of independent variables for GLS estimation. Then, we can obtain a set of OLS estimates μ ^ m o l s P m Y for m = 1 , , M 1 and a set of GLS estimates μ ^ m g l s G m Y for m = 1 , , M 2 . Therein, P m = X m ( X m X m ) 1 X m is the projection matrix of the m th regression model for OLS with m = 1 , , M 1 , and G m X m ( X m Ω 1 X m ) 1 X m Ω 1 with X m being the independent variable matrix of the m th regression model for GLS with m = 1 , , M 2 . In this paper, we only consider the situation with nested models for both OLS and GLS estimators. This means that the m th model is nested in the ( m + 1 ) th model. Theoretical results may be extended to non-nested candidate models using the approach of Wan et al. (2010). Moreover, M 1 and M 2 can be fixed or go to infinity when the sample size n is increasing.
Based on those OLS and GLS estimates, we construct a combination estimator as follows:
μ ^ W = m = 1 M 1 w m o l s μ ^ m o l s + m = 1 M 2 w m g l s μ ^ m g l s , μ ^ o l s W 1 + μ ^ g l s W 2
where W = W 1 , W 2 = w 1 o l s , , w M 1 o l s , w 1 g l s , , w M 2 g l s belongs to:
H = W W 0 , 1 M , I W = 1 ,
where M = M 1 + M 2 and I denote a M × 1 vector having all elements equal to one.
In order to reduce the risk of the combination-estimation method proposed above, we need to select a suitable weight vector. To do that, in this subsection, we propose two versions, MMA-GLSMA and HRCp-GLSMA, with infeasible criteria. In the next subsection, we provide their feasible counterparts for the situation with unknown Ω values.
HRCp-GLSMA: The first information criterion for selecting a weight vector is defined as:
C ¯ n W = s 1 2 Y μ ^ o l s W 1 * 2 + 2 t r P W 1 * Ω
+ s 2 2 Y μ ^ g l s W 2 * 2 + 2 t r G W 2 * Ω
+ 2 s 1 s 2 Y μ ^ o l s W 1 * Y μ ^ g l s W 2 *
+ t r P W 1 * Ω + t r G W 2 * Ω e e ,
where s 1 m = 1 M 1 w m o l s , s 2 m = 1 M 2 w m g l s , W 1 * W 1 / s 1 , W 2 * W 2 / s 2 , P W 1 * m = 1 M 1 w m o l s P m and G W 2 * m = 1 M 2 w m g l s G m .
Note that:
C ¯ n W = s 1 2 H R C p W 1 * + s 2 2 C I n W 2 * + 2 s 1 s 2 Y μ ^ o l s W 1 * Y μ ^ g l s W 2 * + t r P W 1 * Ω + t r G W 2 * Ω e e ,
where H R C p W 1 * = Y μ ^ o l s W 1 * 2 + 2 t r P W 1 * Ω is the HRCp model-averaging criterion proposed by Liu and Okui (2013) with the weight vector W 1 * and C I n W 2 * = Y μ ^ g l s W 2 * 2 + 2 t r G W 2 * Ω is a GLSMA information criterion proposed by Liu et al. (2016) with the weight vector W 2 * . C ¯ can be regarded as a combination of the HRCp and the GLSMA. Hence, we call C ¯ the HRCp-GLSMA-type criterion.
MMA-GLSMA: Second, we propose an MMA-GLSMA-type criterion for weight selection. The infeasible MMA-GLSMA-type criterion is defined as follows:
C ˜ n W = s 1 2 C M M A W 1 * + s 2 2 C I n W 2 * + 2 s 1 s 2 Y μ ^ o l s W 1 * Y μ ^ g l s W 2 * + σ 2 t r P W 1 * + t r G W 2 * Ω e e ,
where C M M A W 1 * = Y μ ^ o l s W 1 * 2 + 2 σ 2 t r P W 1 * is the MMA criterion proposed by Hansen (2007) with the weight vector W 1 * .
Suppose the variance-covariance matrix Ω is known; we can then choose the weight by minimizing the criteria C ¯ n or C ˜ n , as follows:
W ¯ = arg min W H C ¯ n
or
W ˜ = arg min W H C ˜ n .
However, since the variance-covariance matrix Ω is unknown, W ¯ and W ˜ are infeasible.

2.2. Feasible Combination Estimator and Information Criteria

For a situation with unknown variance, a feasible combination estimator can be constructed using feasible GLS (FGLS) estimators. FGLS estimators and the feasible combination estimator are defined below:
μ ^ F W = m = 1 M 1 w m o l s μ ^ m o l s + m = 1 M 2 w m g l s μ ^ m f g l s , μ ^ o l s W 1 + μ ^ f g l s W 2
where the FGLS estimator is μ ^ m f g l s X m ( X m Ω ^ 1 X m ) 1 X m Ω ^ 1 Y . Therein, the estimator Ω ^ is based on the k-NN estimator σ i ˇ 2 of Liu et al. (2016).
We propose two feasible counterparts of C ¯ n and C ˜ n . The feasible HRCp-GLSMA-type criterion C ¯ n F is defined as:
C ¯ n F W = s 1 2 Y μ ^ o l s W 1 * 2 + 2 i = 1 n e ^ i 2 p i i W 1 * + s 2 2 Y μ ^ f g l s W 2 * 2 + 2 t r G F W 2 * Ω ^ + 2 s 1 s 2 Y μ ^ o l s W 1 * Y μ ^ f g l s W 2 * + t r P W 1 * Ω ^ + t r G F W 2 * Ω ^ e ^ e ^ ,
and the feasible MMA-GLSMA-type criterion C ˜ n F is defined as:
C ˜ n F W = s 1 2 Y μ ^ o l s W 1 * 2 + 2 σ ^ 2 t r P W 1 * + s 2 2 Y μ ^ f g l s W 2 * 2 + 2 t r G F W 2 * Ω ^ + 2 s 1 s 2 Y μ ^ o l s W 1 * Y μ ^ f g l s W 2 * + σ ^ 2 t r P W 1 * + t r G F W 2 * Ω ^ e ^ e ^ ,
where e ^ i denotes the i th element of e ^ n / n k L I P L Y with k L denoting the number of independent variables in the largest OLS model:
G F W 2 * = m = 1 M 2 w m g l s X m ( X m Ω ^ X m ) 1 X m Ω ^ 1 .
P L denotes the projection matrix of that model, and p i i denotes the i th diagonal element of P W 1 * . e ˜ I P L Y , while σ ^ 2 n k L 1 e ˜ e ˜ is defined as one of the estimators suggested by Hansen (2007). Ω ^ is calculated by plugging in the k-NN estimator σ i ˇ 2 used in Liu et al. (2016).

3. Properties of the Criteria

The following lemma demonstrates the significant fact that the two infeasible criteria proposed above are unbiased estimates of the risk function R W E L W plus a constant; L W μ μ ^ W 2 is the loss function.
Lemma 1.
For any real-valued vector W, E C ¯ n W = R W + c 1 and E C ˜ n W = R W + c 2 , where c 1 and c 2 are constants.
Another useful property is that all criteria are asymptotically optimal in the sense of Li (1987). Proofs of the asymptotic optimality for all of the above-mentioned criteria can be performed by extending the proofs of Hansen (2007); Liu and Okui (2013); Liu et al. (2016). As an example, we demonstrate the optimality of the feasible MMA-GLSMA case method with the nonparametric estimator of Ω used in Liu et al. (2016).
To do that, we define the feasible loss function and the risk function as follows:
L F W μ μ ^ F ( W ) 2 = ( μ ^ F ( W ) μ ) ( μ ^ F ( W ) μ ) .
We employ some notations and assumptions from Liu et al. (2016), reproduced in the Appendix, and add the following additional assumption:
Assumption 1.
As n and k l , k l + k l j = 1 n b l j 2 / ξ n 0 , where ξ n inf W H n ( N ) R W , k l is the number of regressors used in the regression model for the k-NN estimator and b l i denotes the approximation error of that model.
Assumption 1’ guarantees that when the number of regressors used in the regression model (adjusted with the approximation errors of that model) increases with the sample size n, it increases at a rate slower than the lower bound of the risk across all possible weights. In practice, this assumption requires us to moderate the increase in the number of regressors k l (relative to the sample size) to reduce the approximation errors.
Following Hansen (2007), we restrict the elements of the weighting vector to belong to set H M ( N ) { 0 , 1 / N , 2 / N , , 1 } for some integer N < .
Theorem 1.
Under Assumption 1’ and Assumptions 1–3, 6, and 10–14 of Liu et al. (2016), as n , k l , κ and k L / n 0 , we have:
L F ( W ˜ ) inf W H M ( N ) L ( W ) p 1 ,
where W ˜ = arg min W H M ( N ) C ˜ n F W .
In other words, as the sample size increases, our method can achieve the infimum of the loss.

4. Simulation Study

To investigate the finite sample performance of the proposed MMA-GLSMA and HRCp-GLSMA versions and to compare them with other alternative methods, we performed a Monte Carlo simulation. Alternative methods include MMA, H R C p and GLSMA with C I n F , where C I n F is the feasible counterpart of C I n proposed in Liu et al. (2016).
The data-generating process (DGP) is:
y i = μ i + e i ,
where:
μ i = j = 1 10 , 000 θ j x i j
for i = 1 , , n , with n = 50 . We used parameter θ j = c j 1 with j truncated at 10,000 and a positive constant c. x i 1 = 1 and x i j N 0 , 1 for j 1 are independent with respect to i. We conducted three simulations. The first case was a simulation with a homoscedastic error term e i N 0 , 1 . The second case was a mild heteroscedastic case with heteroscedastic error term e i N 0 , σ i 2 , with σ i = x i 2 . The third case was a strong heteroscedastic case with σ i = x i 2 2 . In all cases, e i was independent with respect to i. The number of regressors in the largest approximation model or the number of nested models was M = 10 . The m th model contained the first m regressors, including the constant term. We varied c to change the R 2 of the DGP from 0 . 1 0 . 9 with an increment of 0 . 1 .
We considered two cases of GLSMA with different estimation methods of σ i , one based on the maximum likelihood estimation (MLE), and the other based on the nonparametric method k-NN. For details of these two cases, see Liu et al. (2016). Because the true specification of σ i is usually unknown in practice, we misspecified σ i for GLSMA. For the MLE-based method, we set σ i 2 = a + b x i 3 4 , where a and b are the unknown parameters to be estimated. For the nonparametric case, we only used x i 3 and x i 4 for k-NN estimation.
The number of replications for all simulations was 1000. We evaluated the performance of each method by the sample mean squared error (MSE) = 1 / 1000 k = 1 1000 μ ^ k μ k 2 , where μ ^ k and μ k are the realized vector of the estimated value μ ^ and true value μ in the k th replication, respectively. The simulation results are shown in Table 1, Table 2 and Table 3. Figure 1 presents the same information relative to the MSE of the GLSMA method.
The results in Table 1 and Table 2 show that our combination methods, MMA-GLSMA and HRCp- GLSMA, performed better than the alternatives (GLSMA, HRCp and MMA) when the error term was homoscedastic or had mild heteroscedasticity for R 2 0 . 7 . When R 2 0 . 8 , the performance of our methods was slightly worse than that of the alternative methods. For the homoscedastic case, the three alternative methods performed similarly. On the other hand, in the case of mild heteroscedasticity, GLSMA and H R C p performed better than MMA (which was expected, as MMA was designed for homoscedastic models).
Table 3 demonstrates that, when the heteroscedasticity of the error term was considerably strong, our combination method HRCp-GLSMA worked much better than the others when the MLE-based estimation of σ i was used. However, MMA-GLSMA and HRCp-GLSMA became worse than GLSMA when σ i was estimated using the nonparametric method.
Moreover, for most cases, GLSMA with the nonparametric estimator of σ i outperformed GLSMA with the MLE-based estimator of σ i . This can be explained by a characteristic of the k-NN method that we adopted. In the k-NN estimation, a large weight was placed on the i th squared residual to estimate σ i 2 ; therefore, even though σ i 2 was misspecified, the estimate could catch the heteroscedasticity of the error term to some extent.
The aforementioned simulation results gave us the following indications for practical analysis. If we know that the heteroscedasticity of the data is considerably strong or the population R 2 is large, we should use the nonparametric GLSMA. Otherwise, it is preferable to choose an MMA-GLSMA or HRCp-GLSMA combination.
Table 4 and Table 5 give the averages of the estimated weights corresponding to the OLS and GLS parts for HRCp-GLSMA and MMA-GLSMA. Table 4 shows that for HRCp-GLSMA, when the heteroscedasticity became stronger, the average weights corresponding to the GLS part increased and the average weights corresponding to the OLS estimates decreased. Table 5 does not show a similar trend for MMA-GLSMA. This might be an explanation for why the performance of MMA-GLSMA was worse than that of HRCp-GLSMA.

5. Conclusions

In this paper, we proposed a combination method of OLS and GLS estimators. The proposed method reduced the risk of misspecification between homoscedastic and heteroscedastic models. The optimality of the criteria for choosing a weight vector was proven under some regularity conditions. We performed simulation experiments to investigate the finite sample property of our combination method. The results of the simulations demonstrate that our method was adaptive for homoscedasticity and heteroscedasticity. As mentioned previously, the proposed method was novel in that it combined estimators from two different estimation methods. Combining the different estimation methods can also combine the advantages of each. This idea could be useful and should be extended to the combination of other estimation methods.

Author Contributions

This is a collaborative project. The authors are listed in alphabetical order. The proofs are done by Q. Liu.

Funding

Q. Liu acknowledges financial support from the JSPS Grant-in-Aid for Young Scientists (B) No. 25780148 and JSPS KAKENHI Grant (C) No. JP16K03590 and No. JP19K01582.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

For the convenience of the readers of this journal, we list Assumptions 1–3, 6 and 10–14 and replicate Lemma 7 of Liu et al. (2016) here. Their notation of R I n ( W ) coincides with our notation R ( W ) ; x m i denotes the i th observation vector of the regressors of the m th model, and x m , j , i is the j th entry of x m i .
Assumption 1.
E ( | e i | 4 ( N + 1 ) ) c < for some c.
Assumption 2.
ξ n inf W H n ( N ) R I n ( W ) as n .
Assumption 3.
0 < inf i σ i 2 sup i σ i 2 < .
Assumption 6.
The maximum eigenvalue of i = 1 n x m i x m i / n is bounded uniformly in n and m. There exists a c > 0 such that the minimum eigenvalue of i = 1 n x m i x m i / n is greater than c for any n and m.
Assume that σ i 2 is a function of a finite subset of x i , denoted by z i , so that σ i 2 = σ 2 ( z i ) .
Assumption 10.
σ 2 ( · ) is differentiable. Let σ ( · ) denote its first derivative. Then, sup z σ ( z ) < . Moreover, the support of z i is bounded, and the density of z i is bounded from below.
Let κ be the tuning parameter for the k-NN estimator. Let c i , 1 i n be such that c i > 0 for 1 i κ , c i = 0 for i > κ and i = 1 κ c i = 1 .
Assumption 11.
lim ¯ n max 1 i κ κ c i < .
Assumption 12.
lim n i = 1 n x m , j , i 4 / n is bounded uniformly in m and j.
Assumption 13.
There exists a C < such that lim n i = 1 n μ i 4 / n < C .
Assumption 14.
Define ν = 4 ( N + 1 ) and:
A n = n 2 / ν κ + k l κ + k l j = 1 n b l j 2 κ + κ 1 / q n 1 / q .
As n , k l and κ , it is the case that k M A n 0 , k M 2 A n / ξ n 0 , n k M A n 2 / ξ n 0 and n k M A n / ξ n 0 .
Lemma 7
of Liu et al. (2016).Suppose that Assumptions 1, 3, 5, 6 and 11 of Liu, Okui and Yoshimura (2016) hold. Then, as n , k l and κ , with k l 2 / κ 0 and k l 2 j = 1 n b l j 2 / κ 0 , we have:
max i | σ ˇ i 2 σ ˜ i 2 | = O p k l κ 1 + j = 1 n b l j 2 .
Proof of Lemma 1.
We have:
R W = E μ μ ^ W 2 = E s 1 μ s 1 μ ^ o l s W 1 * + s 2 μ s 2 μ ^ g l s W 2 * 2
= E s 1 2 μ μ ^ o l s W 1 * 2
+ E 2 s 1 s 2 μ μ ^ o l s W 1 * μ μ ^ g l s W 2 *
+ E s 2 2 μ μ ^ g l s W 2 * 2 .
According to Liu and Okui (2013); Liu et al. (2016), it is obvious that the squared terms (1) and (2) are unbiased estimates of the terms (A1) and (A3) plus a constant.
In order to estimate the cross-product term (A2), we can observe that:
Y μ ^ o l s W 1 * Y μ ^ g l s W 2 * = μ μ ^ o l s W 1 * μ μ ^ g l s W 2 * + e e + e , I P W 1 * μ + e , I G W 2 * μ e P W 1 * e e G W 2 * e
and E e , I P W 1 * μ = E e , I G W * 2 μ = 0 . Therefore, the sum of the terms (3) and (4) unbiasedly estimates the cross-product term (A2). □
Proof of Theorem 1.
We define L o l s W μ μ ^ o l s W μ μ ^ o l s W and L f g l s W μ μ ^ f g l s W μ μ ^ f g l s W . Note that:
C ˜ n F W L F ( W ) = s 1 2 C M M A F W 1 * L o l s W 1 * + s 2 2 C I n F W 2 * L f g l s W 2 * + 2 s 1 s 2 e , I P W 1 * μ + e , I G F W 2 * μ + σ ^ 2 t r P W 1 * + t r G F W 2 * Ω ^ e P W 1 * e e G F W 2 * e + e e e ^ e ^ ,
where C M M A F = C M M A W 1 * = Y μ ^ o l s W 1 * 2 + 2 σ ^ 2 t r P W 1 * and C I n F W 2 * = Y μ ^ g l s W 2 * 2 + 2 t r G F W 2 * Ω ^ .
Using the results of Hansen (2007); Liu et al. (2016), it can be shown that the supremum with respect to W H M ( N ) of all the absolute values of the terms except e e e ^ e ^ divided by risk function R W go to zero in probability. We only need to show:
sup W e e e ^ e ^ R W p 0 .
This can easily be done by modifying Lemma 7 of Liu et al. (2016). By replacing s i j , the weight defined for k-NN in Liu et al. (2016) with one, we have:
sup W e e e ^ e ^ R W e e e ˜ e ˜ / ξ n + k L / n k L e ˜ e ˜ / ξ n = O p k l + k l j = 1 n b l j 2 / ξ n + o p 1 0 .
The proof is complete.  □

References

  1. Danilov, Dmitry, and Jan R. Magnus. 2004. On The Harm That Ignoring Pretesting Can Cause. Journal of Econometrics 122: 27–46. [Google Scholar] [CrossRef]
  2. Hansen, Bruce E. 2007. Least Squares Model Averaging. Econometrica 75: 1175–89. [Google Scholar] [CrossRef]
  3. Hansen, Bruce E., and Jeffrey S. Racine. 2012. Jackknife Model Averaging. Journal of Econometrics 167: 38–46. [Google Scholar] [CrossRef]
  4. Li, Ker-Chau. 1987. Asymptotic Optimality for Cp, CL, Cross-Validation and Generalized Cross-Validation: Discrete Index Set. Annals of Statistics 15: 958–75. [Google Scholar] [CrossRef]
  5. Liu, Qingfeng, and Ryo Okui. 2013. Heteroscedasticity-Robust Cp Model Averaging. Econometrics Journal 16: 463–72. [Google Scholar] [CrossRef]
  6. Liu, Qingfeng, Ryo Okui, and Arihiro Yoshimura. 2016. Generalized Least Squares Model Averaging. Econometric Reviews 35: 1692–752. [Google Scholar] [CrossRef]
  7. Makridakis, Spyros, Evangelos Spiliotis, and Vassilios Assimakopoulos. 2018. The M4 Competition: Results, Findings, Conclusion and Way Forward. International Journal of Forecasting 34: 802–8. [Google Scholar] [CrossRef]
  8. Wan, Alan TK, Xinyu Zhang, and Guohua Zou. 2010. Least Squares Model Averaging by Mallows Criterion. Journal of Econometrics 156: 277–83. [Google Scholar] [CrossRef]
  9. Xie, Tian. 2017. Heteroscedasticity-Robust Model Screening: A Useful Toolkit for Model Averaging in Big Data Analytics. Economics Letters 151: 119–22. [Google Scholar] [CrossRef]
  10. Yuan, Zheng, and Yuhong Yang. 2005. Combining Linear Regression Models: When and How? Journal of the American Statistical Association 100: 1202–14. [Google Scholar] [CrossRef]
  11. Zhang, Xinyu, Alan TK Wan, and Guohua Zou. 2013. Model Averaging by Jackknife Criterion in Models with Dependent Data. Journal of Econometrics 174: 82–94. [Google Scholar] [CrossRef]
  12. Zhang, Xinyu, Guohua Zou, and Raymond J. Carroll. 2015. Model Averaging Based on Kullback-Leibler Distance. Statistica Sinica 25: 1583–98. [Google Scholar] [CrossRef] [PubMed]
  13. Zhao, Shangwei, Aman Ullah, and Xinyu Zhang. 2018. A Class of Model Averaging Estimators. Economics Letters 162: 101–6. [Google Scholar] [CrossRef]
Figure 1. Performance of the methods relative to generalized least-squares model averaging (GLSMA): Heteroscedasticity-robust Cp (HRCp), Mallows model averaging (MMA), MMA-GLSMA and HRCp-GLSMA.
Figure 1. Performance of the methods relative to generalized least-squares model averaging (GLSMA): Heteroscedasticity-robust Cp (HRCp), Mallows model averaging (MMA), MMA-GLSMA and HRCp-GLSMA.
Econometrics 07 00038 g001
Table 1. Sample mean squared error (MSE) for the case with homoscedastic error.
Table 1. Sample mean squared error (MSE) for the case with homoscedastic error.
R 2 GLSMAHRCpMMAMMA-GLSMAHRCp-GLSMA
0 . 1 6.806.556.506.166.18
0 . 2 9.509.349.348.818.83
0 . 3 12.2012.0912.1411.5611.56
0 . 4 15.2515.1615.2514.6314.60
Nonpara. 0 . 5 19.0618.9719.0818.5218.47
0 . 6 24.3324.2124.3423.9223.87
0 . 7 32.6432.4832.6132.4832.39
0 . 8 48.6648.3348.5149.1849.19
0 . 9 95.6694.8895.0798.7699.08
0 . 1 6.566.556.506.046.05
0 . 2 9.429.349.348.838.82
0 . 3 12.2212.0912.1411.6111.60
0 . 4 15.3115.1615.2514.7414.67
MLE 0 . 5 19.1418.9719.0818.6318.56
0 . 6 24.4424.2124.3424.0023.95
0 . 7 32.7432.4832.6132.5032.43
0 . 8 48.7248.3348.5149.1349.04
0 . 9 95.4294.8895.0798.7098.56
Table 2. Sample MSE for the case with mild heteroscedastic error.
Table 2. Sample MSE for the case with mild heteroscedastic error.
R 2 GLSMAHRCpMMAMMA-GLSMAHRCp-GLSMA
0 . 1 6.436.516.976.186.00
0 . 2 8.629.139.378.328.18
0 . 3 10.7711.5811.7410.4810.36
0 . 4 13.0214.0714.2112.8012.77
Nonpara. 0 . 5 15.7616.9417.1115.6515.61
0 . 6 19.4520.6620.8619.4519.36
0 . 7 25.2026.3426.5625.2925.28
0 . 8 36.0537.0437.2636.5036.44
0 . 9 67.4667.8768.1069.5369.34
0 . 1 6.806.516.976.476.30
0 . 2 9.239.139.378.738.58
0 . 3 11.5911.5811.7411.0210.90
0 . 4 14.0614.0714.2113.4813.39
MLE 0 . 5 16.9516.9417.1116.4116.32
0 . 6 20.7120.6620.8620.2520.14
0 . 7 26.4326.3426.5626.0425.96
0 . 8 37.2137.0437.2637.1337.07
0 . 9 68.3667.8768.1069.8669.79
Table 3. Sample MSE for the case with strong heteroscedastic error.
Table 3. Sample MSE for the case with strong heteroscedastic error.
R 2 GLSMAHRCpMMAMMA-GLSMAHRCp-GLSMA
0 . 1 13.4414.8820.0715.5213.70
0 . 2 16.1318.6823.1718.0516.30
0 . 3 18.9222.7426.4720.8019.17
0 . 4 22.0227.1130.2424.0622.69
Nonpara. 0 . 5 25.9132.1834.7427.9626.73
0 . 6 30.8938.5240.4833.1332.21
0 . 7 38.2847.0348.4640.6239.94
0 . 8 51.1560.6861.9253.8053.37
0 . 9 85.6594.9796.0689.2189.15
0 . 1 17.0214.8820.0717.5215.86
0 . 2 20.1518.6823.1720.2318.64
0 . 3 23.5222.7426.4723.4021.94
0 . 4 27.4027.1130.2426.9925.59
MLE 0 . 5 32.0132.1834.7431.4030.21
0 . 6 37.9038.5240.4837.0236.00
0 . 7 46.3547.0348.4645.1344.29
0 . 8 59.5660.6861.9258.6557.93
0 . 9 93.8694.9796.0693.6593.33
Table 4. Averages of the OLS ( W ¯ 1 ) and GLS ( W ¯ 2 ) parts of the weight vector for HRCp-GLSMA.
Table 4. Averages of the OLS ( W ¯ 1 ) and GLS ( W ¯ 2 ) parts of the weight vector for HRCp-GLSMA.
R 2 0.10.20.30.40.50.60.70.80.9
Homoscedastic Cases
Nonp.OLS ( W ¯ 1 )0.500.500.500.500.490.490.490.490.49
Nonp. GLS ( W ¯ 2 )0.500.500.500.500.510.510.510.510.51
MLE OLS ( W ¯ 1 )0.500.500.500.500.500.500.500.500.50
MLE GLS ( W ¯ 2 )0.500.500.500.500.500.500.500.500.50
Mild Heteroscedastic Cases
Nonp. OLS ( W ¯ 1 )0.490.490.490.490.490.490.490.490.49
Nonp. GLS ( W ¯ 2 )0.510.510.510.510.510.510.510.510.51
MLE OLS ( W ¯ 1 )0.490.490.490.490.490.490.500.500.50
MLE GLS ( W ¯ 2 )0.510.510.510.510.510.510.500.500.50
Strong Heteroscedastic Cases
Nonp. OLS ( W ¯ 1 )0.480.480.480.480.480.480.480.480.48
Nonp. GLS ( W ¯ 2 )0.520.520.520.520.520.520.520.520.52
MLE OLS ( W ¯ 1 )0.490.490.490.490.490.490.490.490.49
MLE GLS ( W ¯ 2 )0.510.510.510.510.510.510.510.510.51
Table 5. Averages of the OLS ( W ¯ 1 ) and GLS ( W ¯ 2 ) parts of the weight vector for MMA-GLSMA.
Table 5. Averages of the OLS ( W ¯ 1 ) and GLS ( W ¯ 2 ) parts of the weight vector for MMA-GLSMA.
R 2 0.10.20.30.40.50.60.70.80.9
Homoscedastic Cases
Nonp. OLS ( W ¯ 1 )0.500.490.490.490.490.490.490.490.49
Nonp. GLS ( W ¯ 2 )0.500.510.510.510.510.510.510.510.51
MLE OLS ( W ¯ 1 )0.500.500.500.500.500.500.500.500.50
MLE GLS ( W ¯ 2 )0.500.500.500.500.500.500.500.500.50
Mild Heteroscedastic Cases
Nonp. OLS ( W ¯ 1 )0.500.490.490.490.490.490.490.490.49
Nonp. GLS ( W ¯ 2 )0.500.510.510.510.510.510.510.510.51
MLE OLS ( W ¯ 1 )0.500.500.500.500.500.500.500.500.50
MLE GLS ( W ¯ 2 )0.500.500.500.500.500.500.500.500.50
Strong Heteroscedastic Cases
Nonp. OLS ( W ¯ 1 )0.500.500.490.490.490.490.490.490.49
Nonp. GLS ( W ¯ 2 )0.500.500.510.510.510.510.510.510.51
MLE OLS ( W ¯ 1 )0.500.500.500.500.500.500.500.500.50
MLE GLS ( W ¯ 2 )0.500.500.500.500.500.500.500.500.50

Share and Cite

MDPI and ACS Style

Liu, Q.; Vasnev, A.L. A Combination Method for Averaging OLS and GLS Estimators. Econometrics 2019, 7, 38. https://0-doi-org.brum.beds.ac.uk/10.3390/econometrics7030038

AMA Style

Liu Q, Vasnev AL. A Combination Method for Averaging OLS and GLS Estimators. Econometrics. 2019; 7(3):38. https://0-doi-org.brum.beds.ac.uk/10.3390/econometrics7030038

Chicago/Turabian Style

Liu, Qingfeng, and Andrey L. Vasnev. 2019. "A Combination Method for Averaging OLS and GLS Estimators" Econometrics 7, no. 3: 38. https://0-doi-org.brum.beds.ac.uk/10.3390/econometrics7030038

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop