Original Article | OPEN ACCESS DOI: 10.23937/2469-5831/1510031

A New Two Parameter Biased Estimator for the Unrestricted Linear Regression Model: Theory, Simulation and Application

Issam Dawoud1 and BM Golam Kibria2*

1Department of Mathematics, Al-Aqsa University, Gaza, Palestine

2Department of Mathematics and Statistics, Florida International University, Miami, Florida, USA

Abstract

This paper proposes a new biased estimator for estimating the regression parameters for the multiple linear regression models when the regressors are correlated. Theoretical comparisons and simulations results show that, the proposed estimator performs better than other existing estimators under some conditions in the smaller mean squares error sense. A real-life dataset is analyzed to illustrate the findings of the paper.

Keywords

Biased estimator, Liu estimator, Monte Carlo simulation, Multicollinearity, MSE, Ridge regression estimator

Introduction

The ordinary least squares (OLS) estimator is the best linear unbiased estimator and has been used to estimate the parameters of the linear regression model since its inception. One of the important assumptions for the linear regression model is that the regressors (independent variables) are independent. However, in practice the regressors may or may not be independent, which causes the problem of multicollinearity. In the presence of the multicollinearity, the OLS estimator is inefficient and gives wrong sign of the parameters in the multiple linear regression models [1]. To handle these problems, many authors have given different types of estimators, to mention a few, [1-7], and recently [8], for one parameter biased estimator. However, to refer for two parameters mode, the following authors are notable, [9-15], among others.

The main objective of this paper is to propose a new two parameter biased estimator for the regression coefficients and then to compare the performance of the new estimator with the OLS, the ordinary ridge regression (ORR) of Hoerl and Kennard [1], the Liu of Liu [6] and the Kibria-Lukman (KL) of Kibria and Lukman [8] estimators. The rest of the paper is organized as follows: Some estimators and their statistical properties are given in Section 2. In section 3, the theoretical comparisons among the proposed estimator and existing estimators and the biasing parameters k and d are given. A Monte Carlo simulation study is performed in section 4. A real-life data are analyzed in section 5. Finally, Some Conclusions are given in section 6.

The Model and Estimators

Consider the following linear regression model:

y=Xβ+ε,          (2.1)

where y is an n×1 vector of the response variable, X is a known n×p full rank matrix of the explanatory variables, β is an p×1 vector of unknown regression coefficients, ε is a n × 1 vector of disturbance assumed to be distributed with mean vector 0 and variance covariance matrix σ2I, and I is an identity matrix of order n × n. To define various estimators, canonical form of the model (2.1) is given by:

y=Zα+ε, ε~N(0, σ 2 I n )            (2.2)

where, Z=XC, α= C β , and C is an orthogonal matrix such that Z Z= C X XC=U=diag( u 1 , u 2 ,, u p ). Then, the OLS estimator of α is given as:

α ^ = U 1 Z y,       (2.3)

and then the mean squared error matrix (MSEM) of α ^ is given by

MSEM( α ^ )= σ 2 U 1 .            (2.4)

The ORR of α [1] is

α ^ k =NU α ^ ,          (2.5)

Where, N=[U+kI]-1 and the MSEM of α ^ k would be

MSEM( α ^ k )= σ 2 NU N +(NU I p )α α (NU I p )               (2.6)

Hoerl, et al. [16] defined the biasing parameter k for α ^ k as follows:

k ^ HM = p σ ^ 2 i=1 p α ^ i 2 .                 (2.7)

The Liu estimator of α [6] is

α ^ d =F α ^ ,           (2.8)

Where, F= [U+ I p ] 1 [U+d I p ] and the biasing parameter d of α ^ d is given by

d ^ opt =1 σ ^ 2 i=1 p 1/( u i ( u i +1)) i=1 p ( α ^ i 2 / ( u i +1) 2 )               (2.9)

MSEM( α ^ d )= σ 2 F U 1 F + (1d) 2 (U+ I p ) 1 α α (U+ I p ) 1 .               (2.10)

If d ^ opt is negative, Ozkale and Kaciranlar [9] adopt the alternative biasing parameter below:

d ^ alt =min α ^ i 2 ( σ ^ 2 / u i )+ α ^ i 2 i=1 p .               (2.11)

The recently proposed KL estimator by Kibria and Lukman [8] is defined as,

α ^ KL =NM α ^ ,            (2.12)

Where, M=(Uk Ι p ) and the biasing parameter k of the KL estimator is defined by

k ^ min (KL)=min σ ^ 2 2 α ^ i 2 +( σ ^ 2 / u i )                   (2.13)

MSEM( α ^ KL )= σ 2 NM U 1 M N +[NM I p ]α α [NM I p ]           (2.14)

The new biased (NB) estimator of α is obtained by minimizing (yZα ) (yZα) subject to

( k α+ k (1+d) α ^ ) ( k α+ k (1+d) α ^ )=c where c is a constant,

(yZα ) (yZα)+[( k α+ k (1+d) α ^ ) ( k α+ k (1+d) α ^ )c].               (2.15)

Here, k and d are the Lagrangian multipliers.

The solution to (2.15) gives the new estimator as follows:

α ^ NB = (U+k Ι p ) 1 (Uk(1+d) Ι p ) α ^ =NB α ^                 (2.16)

where B=(Uk(1+d) Ι p ),  k0 and 0<d<1.

Moreover, the proposed NB estimator is also obtained by augmenting k (1+d) α ^ = k α+ ε to equation (2.2) and then using the OLS estimate.

The proposed NB estimator is a general estimator which includes the OLS and the KL estimator:

1 If k=0 in α ^ NB becomes α ^ .

2 If d=0 in α ^ NB becomes α ^ KL .

The MSEM of the proposed NB estimator of α is given by

MSEM( α ^ NB )= σ 2 NB U 1 B N +[NB I p ]α α [NB I p ] .                (2.17)

The lemmas below will be used for theoretical comparisons among estimators in the next section.

Lemma 2.1: [17] Let E be an n×n positive definite matrix, that is E>0 and α be some vector; then, Eα α >0 if and only if α E 1 α<1.

Lemma 2.2: [18] Let α i = A i y, i=1,2 be two linear estimators of α. Suppose that D=Cov( α ^ 1 )Cov( α ^ 2 )>0, where Cov( α ^ i )i=1,2 be the covariance matrix of α ^ i and b i =Bias( α ^ i )=( A i XI)α, i = 1,2. Consequently,

Δ( α ^ 1 α ^ 2 )=MSEM( α ^ 1 )MSEM( α ^ 2 )= σ 2 D+ b 1 b 1 b 2 b 2 >0              (2.18)

if and only if b 2 [ σ 2 D+ b 1 b 1 ] 1 b 2 <1 where MSEM( α ^ i )=Cov( α ^ i )+ b i b i .

Comparison among the Estimators

Comparison between α ^ and α ^ NB

Theorem 3.1: MSEM( α ^ )MSEM( α ^ NB )>0 if and only if

α [NB I p ] [ σ 2 ( U 1 NB U 1 B N )][NB I p ]α<1          (3.1)

Proof:

D( α ^ )D( α ^ NB )= σ 2 ( U 1 NB U 1 B N ) = σ 2 diag 1 u i ( u i k(1+d)) 2 u i ( u i +k) 2 i=1 p (3.2)

Where U 1 NB U 1 B N will be positive definite (pd) if and only if ( u i +k) 2 ( u i k(1+d)) 2 >0 or ( u i +k)( u i k(1+d))>0. Clearly, for k>0 and 0<d<1, ( u i +k)( u i k(1+d))=k(2+d)>0. By Lemma 2.2. The proof is completed.

Comparison between α ^ k and α ^ NB

Theorem 3.2: MSEM( α ^ k )MSEM( α ^ NB )>0 if and only if

α [NB I p ] [ V 1 +(NU I p )α α (NU I p ) ][NB I p ]α<1                 (3.3)

Where, V 1 = σ 2 (NU N NB U 1 B N )

Proof:

V 1 = σ 2 (NU N NB U 1 B N ) = σ 2 diag u i ( u i +k) 2 ( u i k(1+d)) 2 u i ( u i +k) 2 i=1 p (3.4)

Where, NU N NB U 1 B N will be pd if and only if u i 2 ( u i k(1+d)) 2 >0 or u i ( u i k(1+d))>0. Clearly, for k>0 and 0<d<1 , u i ( u i k(1+d))=k(1+d)>0. By Lemma 2.2. The proof is completed.

Comparison between α ^ d and α ^ NB

Theorem 3.3: MSEM( α ^ d )MSEM( α ^ NB )>0 if and only if

α [NB I p ] [ V 2 +(F I p )α α (F I p ) ][NB I p ]α<1             (3.5)

where, V 2 = σ 2 (F U 1 FNB U 1 B N )

Proof:

V 2 = σ 2 (F U 1 FNB U 1 B N ) = σ 2 diag ( u i +d) 2 u i ( u i +1) 2 ( u i k(1+d)) 2 u i ( u i +k) 2 i=1 p (3.6)

Where, F U 1 FNB U 1 B N will be pd if and only if ( u i +d) 2 ( u i +k) 2 ( u i +1) 2 ( u i k(1+d)) 2 >0 or ( u i +d)( u i +k)( u i +1)( u i k(1+d))>0. Clearly, for k>0 and 0<d<1, ( u i +d)( u i +k)( u i +1)( u i k(1+d))= u i (2k+kd+d1)+k(2d+1)>0. By Lemma 2.2. The proof is completed.

Comparison between α ^ KL and α ^ NB

Theorem 3.4: MSEM( α ^ KL )MSEM( α ^ NB )>0 if and only if

α [NB I p ] [ V 3 +(NM I p )α α (NM I p ) ][NB I p ]α<1                    (3.7)

Where, V 3 = σ 2 (NM U 1 M N NB U 1 B N )

Proof:

V 2 = σ 2 (NM U 1 M N NB U 1 B N ) = σ 2 diag ( u i k) 2 u i ( u i +k) 2 ( u i k(1+d)) 2 u i ( u i +k) 2 i=1 p (3.8)

Where, NM U 1 M N NB U 1 B N will be pd if and only if ( u i k) 2 ( u i k(1+d)) 2 >0 or ( u i k)( u i k(1+d))>0. Clearly, for k>0 and 0<d<1,  ( u i k)( u i k(1+d))=kd>0. By Lemma 2.2 The proof is completed.

Selection of the parameters k and d

Different biasing parameters estimators of k and d are proposed in different studies, for example, Hoerl and Kennard [1], Liu [6], [16,19-23], among others.

So the optimal values of k and d for the proposed NB estimator is going to be found. At first, by minimizing the equation m, we obtain the optimal value of k when d is fixed as

MSEM( α ^ NB )=E(( α ^ NB α ) ( α ^ NB α)),

m=tr(MSEM( α ^ NB )),

m= σ 2 i=1 p ( u i k(1+d)) 2 u i ( u i +k) 2 + k 2 (2+d) 2 i=1 p α i 2 ( u i +k) 2                 (3.9)

Differentiating m with respect to k and setting ( m/ k )=0, we get

k= u i σ 2 (d+1)( u i α i 2 + σ 2 )+ u i α i 2 .             (3.10)

Then, the estimated optimal value of k is given as follows:

k ^ = u i σ ^ 2 (d+1)( u i α ^ i 2 + σ ^ 2 )+ u i α ^ i 2 ,                (3.11)

and,

k ^ min (NB)=min u i σ ^ 2 (d+1)( u i α ^ i 2 + σ ^ 2 )+ u i α ^ i 2 i=1 p .             (3.12)

Also, the optimal value of d will be found by differentiating m with respect to d when k is fixed and setting ( m/ d )=0, we get

d= u i σ 2 (2 u i α i 2 + σ 2 )k k( u i α i 2 + σ 2 ) .              (3.13)

Then, the estimated optimal d with the unbiased estimators is given by

d ^ = u i σ ^ 2 (2 u i α ^ i 2 + σ ^ 2 ) k ^ k ^ ( u i α ^ i 2 + σ ^ 2 ) ,            (3.14)

And

d ^ min (NB)=min u i σ ^ 2 (2 u i α ^ i 2 + σ ^ 2 ) k ^ min (NB) k ^ min (NB)( u i α ^ i 2 + σ ^ 2 ) i=1 p .                 (3.15)

The parameters k and d estimators selection in α ^ NB are found iteratively as follows:

1. Find an initial estimate of d using d ^ =min( σ ^ 2 α ^ i 2 ).

2. Determine k ^ min (NB) from (3.12) using d ^ in step 1.

3. Estimate d ^ min (NB) in (3.15) by using k ^ min (NB) in step 2.

4. If d ^ min (NB)>1 or d ^ min (NB)<0, use d ^ min (NB)= d ^ .

Simulation Study

A Monte Carlo simulation study is performed to show the performance of the NB estimator over some existing estimators. It contains two parts: (i) Simulation technique and (ii) Results discussion.

Simulation technique

Using the equation below, we generate the explanatory variables (see, Gibbons [24] and Kibria [19]):

x ij = (1 ρ 2 ) 1/2 z ij +ρ z i,p+1 ,i=1,2,...,n,j=1,2,...,p              (4.1)

where z ij are independent standard normal pseudo-random numbers and ρ is the correlation between any two explanatory variables and ρ here has two values 0.9 and 0.99. The n observations for the response variable y are gotten by the following equation:

y i = β 1 x i1 + β 2 x i2 + β 3 x i3 + e i ,i=1,2,,n              (4.2)

where e i are i.i.dN(0, σ 2 ). The values of β are obtained as β β=1 [25]. Also, we choose the values of the biasing parameters of the estimators as k=d=0.1,0.2,,0.9 by following the work of Wichern and Churchill [26] and Kan, et al. [27] is that when k lies between 0 and 1, the ORR estimator performs better. The Monte Carlo simulation study replication is 1000 times for n = 50 and 100 and σ 2 = 1, 25, and 100. For each replicate, we calculate the mean square error (MSE) of the estimators using the next equation:

MSE( α * )= 1 1000 j=1 1000 ( α ij * α i ) ( α ij * α i )            (4.3)

where α ij * is the estimator and α i is the true parameter. The estimated MSEs of the estimators are shown in Table 1, Table 2, Table 3 and Table 4 for (ρ = 0.90 and n = 50), (ρ = 0.99 and n = 50), (ρ = 0.90 and n = 100), and (ρ = 0.99 and n = 100), respectively.

Table 1: Estimated MSE for OLS, ORR, Liu, KL and the proposed NB. View Table 1

Simulation results discussions

From Table 1, Table 2, Table 3 and Table 4, we observed that, when the factors σ and ρ are going to increase, the estimated MSE values are also going to increase, while n is going to increase, the estimated MSE values are going to decrease. Also, the OLS estimator is performing the worst for all cases in presence of the multicollinearity. Moreover, the simulation results show that, the proposed NB estimator is performing better than the other estimators for most of cases. The Liu estimator gives better results in the MSE values when k=d=0.1,0.2 i.e. when the biasing parameters are near to zero. For ρ = 0.9: The condition number (CN) is approximately around 5 and the variance inflation factors (VIFs) are around 4 to 6 such that it is observed that a close agreement of the proposed NB (although better results) with the KL estimator for low values of k = d and a better performance as k = d increases for a fixed value of σ. This improvement increases with the increase of the value of σ and an even better improvements is observed if ρ = 0.99 where the CN and the VIFs become larger and are approximately around 15 and around 36 to 61, respectively. So, the proposed NB estimator works better for the strong correlated explanatory variables. So, the performance of the proposed NB estimator almost depends on the value of ρ, σ, the biasing parameters k and d, and the true parameter. Thus, simulation results are consistent with the theoretical results.

Table 2: Estimated MSE for OLS, ORR, Liu and the proposed NB. View Table 2

Table 3: Estimated MSE for OLS, ORR, Liu, KL and the proposed NB. View Table 3

Table 4: Estimated MSE for OLS, ORR, Liu, KL and the proposed NB. View Table 4

Application

To illustrate the theoretical and simulation results of this paper, we consider a real life data in this section. The Portland cement data was originally adopted by Woods, et al. [28]. This data was also analyzed by many researchers, for examples, [29,30]; Lukman, et al. [14] and recently by Kibria and Lukman [8], among others. And this data is analyzed here to explain the performance of the proposed NB estimator and the other existing estimators. The regression model for this data is given by

y i = β 0 + β 1 X 1 + β 2 X 2 + β 3 X 3 + β 4 X 4 + ε i .           (5.1)

For knowing more details about this data, we refer Woods, et al. [28].

To show the existence of the multicollinearity, different measures are calculated as the VIFs which are 38.50, 254.42, 46.87 and 282.51, Eigenvalues of X'X are 44676.206, 5965.422, 809.952 and 105.419 and the CN of X'X approximately equals 20.58. The VIFs, the eigenvalues and the CN tell us that there is a severe multicollinearity in the data. Also, σ ^ is equal to 2.446. The correlation coefficients matrix of the explanatory variables are presented in Table 5 such that there is a significant and strong relationship among the following explanatory variables: X 1 and X 3 ,  X 2 and X 4 . Then, the estimated parameters and the MSE values of the estimators are presented in Table 6. It appears from Table 6 that the proposed NB estimator is performing the best where it is giving an obvious improvement over all existing estimators and a little improvement over the KL estimator in which this is consistent with the simulation results because σ ^ here is small and not all the explanatory variables are significantly or strongly correlated at the same degree of correlation even though the data has high CN and VIFs.

Table 5: The correlation coefficients matrix of the explanatory variables. View Table 5

Table 6: The results of regression coeffcients and the corresponding MSE values. View Table 6

Some Concluding Remarks

In this paper, we proposed a new biased (NB) estimator for handling the multicollinearity problem in the multiple linear regression models. Some existing estimators are the special case of the proposed estimator. The proposed NB estimator is compared theoretically with the Ordinary least squares (OLS) estimator, the Ordinary ridge regression (ORR) estimator, the Liu estimator and the Kibria-Lukman (KL) estimator, and then the biasing parameters d and k of the NB estimator are derived. A Monte Carlo simulation study is performed for comparing the performance of the OLS, ORR, Liu, KL and the proposed NB estimators. The main finding of this simulation is that the proposed NB estimator performed better than the above mentioned estimators under some conditions. A real-life data is analyzed to support the findings of the paper.

Acknowledgments

Authors are grateful to referees for their valuable comments and suggestions, which certainly improved the quality and presentation of the paper. Author B M Golam Kibria would like to dedicate this paper to his Ph. D. supervisor, Late Professor M Safiul Haq for his guidance and constant inspiration during his stay at The University of Westen Ontario and finally to achieve this present position.

References

  1. Hoerl AE, Kennard RW (1970) Ridge regression: Biased estimation for nonorthogonal problems. Technometrics 12: 55-67.
  2. Stein C (1956) Inadmissibility of the usual estimator for the mean of a multivariate normal distribution. Proceedings of the Third Berkley Symposium on Mathematical Statistics and Probability 1: 197-206.
  3. Massy WF (1965) Principal components regression in exploratory statistical research. Journal of the American Statistical Association 60: 234-256.
  4. Mayer LS, Willke TA (1973) On biased estimation in linear models. Technometrics 15: 497-508.
  5. Swindel BF (1976) Good ridge estimators based on prior information. Communications in Statistics-Theory and Methods 5: 1065-1075.
  6. Liu K (1993) A new class of biased estimate in linear regression. Communication in Statistics-Theory and Methods 22: 393-402.
  7. Akdeniz F, Kaçiranlar S (1995) On the almost unbiased generalized liu estimator and unbiased estimation of the bias and mse. Communications in Statistics-Theory and Methods 24: 1789-1797.
  8. Kibria BMG, Lukman AF (2020) A new ridge-type estimator for the linear regression model: Simulations and applications. Hindawi Scienti?ca.
  9. Ozkale MR, Kaçiranlar S (2007) The restricted and unrestricted two-parameter estimators. Communications in Statistics-Theory and Methods 36: 2707-2725.
  10. Sakallioglu S, Kaçiranlar S (2008) A new biased estimator based on ridge estimation. Statistical Papers 49: 669-689.
  11. Yang H, Chang X (2010) A new two-parameter estimator in linear regression. Communications in Statistics-Theory and Methods 39: 923-934.
  12. Roozbeh M (2018) Optimal QR-based estimation in partially linear regression models with correlated errors using GCV criterion. Computational Statistics & Data Analysis 117: 45-61.
  13. Akdeniz F, Roozbeh M (2019) Generalized difference-based weighted mixed almost unbiased ridge estimator in partially linear models. Statistical Papers 60: 1717-1739.
  14. Lukman AF, Ayinde K, Binuomote S, Clement OA (2019) Modified ridge-type estimator to combat multicollinearity: Application to chemical data. Journal of Chemometrics 33: e3125.
  15. Lukman AF, Ayinde K, Sek SK, Adewuyi E (2019) A modi?ed new two-parameter estimator in a linear regression model. Modelling and Simulation in Engineering.
  16. Hoerl AE, Kannard RW, Baldwin KF (1975) Ridge regression: Some simulations. Communications in Statistics 4: 105-123.
  17. Farebrother RW (1976) Further results on the mean square error of ridge regression. Journal of the Royal Statistical Society: Series B (Methodological) 38: 248-250.
  18. Trenkler G, Toutenburg H (1990) Mean squared error matrix comparisons between biased estimators-an overview of recent results. Statistical Papers 31: 165-179.
  19. Kibria BMG (2003) Performance of some new ridge regression estimators. Communications in Statistics-Simulation and Computation 32: 419-435.
  20. Kibria BMG, Banik S (2016) Some ridge regression estimators and their performances. Journal of Modern Applied Statistical Methods 15: 206-238.
  21. Lukman AF, Ayinde K (2017) Review and classi?cations of the ridge parameter estimation techniques. Hacettepe Journal of Mathematics and Statistics 46: 953-967.
  22. Mansson K, Kibria BMG, Shukur G (2015) Performance of some weighted Liu estimators for logit regression model: An application to Swedish accident data. Communications in Statistics Theory and Methods 44: 363-375.
  23. Khalaf G, Shukur G (2005) Choosing ridge parameter for regression problems. Communications in Statistics Theory and Methods 34: 1177-1182.
  24. Gibbons DG (1981) A simulation study of some ridge estimators. Journal of the American Statistical Association 76: 131-139.
  25. Newhouse JP, Oman SD (1971) An evaluation of ridge estimators. A report prepared for United States air force project RAND.
  26. Wichern DW, Churchill GA (1978) A comparison of ridge estimators. Technometrics 20: 301-311.
  27. Kan B, Alpu O, Yazici B (2013) Robust ridge and robust liu estimator for regression based on the LTS estimator. Journal of Applied Statistics 40: 644-655.
  28. Woods H, Steinour HH, Starke HR (1932) Effect of composition of Portland cement on heat evolved during hardening. Ind Eng Chem 24: 1207-1214.
  29. Kaciranlar S, Sakallioglu S, Akdeniz F, Styan GPH, Werner HJ (1999) A new biased estimator in linear regression and a detailed analysis of the widely-analysed dataset on portland cement. Sankhya: The Indian Journal of Statistics, Series B 61: 443-459.
  30. Li Y, Yang H (2012) A new Liu-type estimator in linear regression model. Statistical Papers 53: 427-437.

Citation

Dawoud I, Kibria BMG (2020) A New Two Parameter Biased Estimator for the Unrestricted Linear Regression Model: Theory, Simulation and Application. Int J Clin Biostat Biom 6:031. doi.org/10.23937/2469-5831/1510031