Original Research | OPEN ACCESS DOI: 10.23937/2469-5831/1510048

A New Modified Liu Ridge-Type Estimator for the Linear Regression Model: Simulation and Application

Olasunkanmi J Oladapo1, Abiola T Owolabi1*, Janet I Idowu1 and Kayode Ayinde2

1Department of Statistics, Ladoke Akintola University of Technology, Ogbomoso, Oyo State, Nigeria

2Department of Statistics, Federal University of Technology, Akure, Nigeria

Abstract

Several efforts have been made to solve the multicollinearity problem, which arises from correlated regressors in the linear regression model. This is because the Ordinary Least Squares (OLS) Estimator becomes inefficient in the presence of multicollinearity. In this paper, a new modified Liu Ridge-Type estimator called Liu-Dawoud-Kibria is proposed in place of the OLS in estimating the parameters of the general linear model. The theoretical comparison and simulation study results show that the proposed estimator outperforms others under some conditions, using the mean squares error criterion. A real-life dataset is used to bolster the findings of the paper.

Keywords

Liu Dawoud-Kibria, Multicollinearity, OLS estimator, Ridge regression estimator, Monte carlo simulation, Mean square error

Introduction

One of the multiple linear regression model assumptions is that the explanatory variables are independent of each other. However, according to Frisch [1], this independence assumption is often violated in real-life situations leading to the Multicollinearity problem. The ordinary least squares (OLS) estimator has been regarded as one of the most important ways of estimating the parameters of the general linear model since it has minimum variance. But, in the presence of Multicollinearity, the OLSE is no longer a good estimator. Multicollinearity is generally agreed to be present if there is an approximately linear relationship (i.e., shared variance) among some predictor variables in the data [2]. The term multicollinearity refers to a situation where there is an exact (or nearly exact) linear relation among two or more explanatory variables [31]. One of the tests to ascertain the presence of multicollinearity in a dataset is the variance inflation factor (VIF). The VIF values ranging between one and five are moderate and represent a medium level of collinearity. Values of more than five are said to be highly collinear, while values above ten are considered to be extreme. Pairwise correlation of two independent variables can also be used to detect if multicollinearity is present in a dataset. An absolute correlation value of 0.7 and above indicates the presence of multicollinearity.

However, many methods have been proposed to solve this problem by various researchers (Hoerl and Kennard [4]; Liu [5]; Kibria [6]; Sakallioglu and Kaciranlar [7]; Baye and Parker [8]; Yang and Chang [9]; Wu and Yang [10]; Dorugade [11]; Lukman, et al. [12]); Lukman, et al. [13]; Kibria and Lukman [14]; Dawoud and Kibria [15]) and recently Owolabi, et al. [16,17]. This study proposes a new two-parameter estimator to circumvent the problem of Multicollinearity. Aside from multicollinearity, a few terms are defined in the table 1 below.

Table 1: Definition of some basic terms. View Table 1

The organization of the paper is as follows. The model, the proposed and existing estimators are given in Section 2. A comparison of the proposed estimator with some existing ones is shown in section 3. The biasing parameter was obtained in Section 4, while a Monte-Carlo simulation and a numeric example were conducted in Section 5. Section 6 contains concluding remarks.

The Models Specifications and New Estimator

The models and existing estimators

Consider the following linear regression model

y = Xβ + ε,       1

where y is a n × 1 vector of response, X is a n × p full column rank matrix, where n and p refer to the sample size and numbers of explanatory variables., β is a p × 1 vector of unknown parameters, ε is a n × 1 vector of errors supposed to be distributed with mean vector 0 and variance-covariance σ 2 I.

Based on the Gauss-Markov theorem, the ordinary least squares estimator (OLSE) is given as:

β ^  =  X'X 1 X'y            2

The canonical form of Eq. (1) is rewritten as:

y=Zα+ε             3

where Z = XQ, α = Q'β and Q is the orthogonal matrix whose columns constitute the eigenvectors of X'X. Then Z'Z = Q'X'XQ = Λ = diag(λ1,..., λp),

where λ1 ≥ λ2 ≥ ...λp > 0 are the ordered eigenvalues X'X. The ordinary least square estimator (OLSE) of equation (2) can be defined as:

α ^  =  Λ 1 X'y            4

And the Mean Square Error matrix (MSEM) of α ^ is defined as

MSEM α ^  =  Λ 1 σ 2          5

The Ordinary Ridge Regression Estimator by Hoerl and Kennard [4] is given as: α ^ k =ΑX'y       6

where A = (Λ + kI)-1 and k is a non-negative biasing parameter. And the MSEM is given as:

MSEM α ^ k  =  σ 2 AΛA +  AΛ - I αα' AΛ - I '       7

The Liu estimator is defined as:

α ^ d  = E α ^    8

where E = (Λ + I)-1(Λ + dI) and d is a biasing parameter of Liu Estimator. The MSEM of α ^ d is defined as:

MSEM α ^ d  =  σ 2 E Λ 1 E +  E - I αα' E - I '       9

The Kibria-Lukman (KL) estimator is defined as:

α ^ KL  = P α ^        10

where P = (Λ + kI)-1(Λ - kI) and the MSEM of α ^ KL is given as

MSEM α ^ KL  =  σ 2 P Λ 1 P +  P - I αα' P - I '         11

The Modified Ridge Type (MRT) estimator is defined as:

α ^ MRT  = H α ^     12

where H = Λ(Λ + k(1 + d)I)-1 and the MSEM is given as

MSEM( α ^ MRT ) =  σ 2 H Λ -1 H +  HI αα'(HI)'             13

The Dawoud Kibria (DK) estimator is defined as:

α ^ DK  = KM α ^             14

where K = (Λ + k(1 + d))-1, M = (Λ - k(1 + d)) and the MSEM is given as MSEM α ^ DK  =  σ 2 KM Λ 1 KM +  KMI αα' KMI '           15

The proposed estimator

In this paper, we proposed a New Biasing Estimator called Liu Dawoud-Kibria Estimator by following a method similar to that proposed by Liu [5]; Kaciranlar, et al. [20]; Yang and Chang [9], and Dawoud, et al. [21]. The proposed New Biasing Liu Dawoud-Kibria estimator for α is obtained by replacing with in the Liu estimator, and it becomes as follows:

α ^ LDK  = WS α ^         16

Where W = (Λ + I)-1(Λ + dI) S = (Λ + k(1 + d)I)-1(Λ – k(1 + d)I), d and k are the biasing parameters.

The properties of the new estimator, namely; the bias vector, covariance, and mean squared error matrix (MSEM) of the proposed estimator, are given as follows:

B α ^ LDK  =  WSI α          17

D α ^ LDK  =  σ 2 WS Λ -1 WS          18

MSEM α ^ LDK  =  σ 2 WS Λ 1 WS + (WSI)αα'(WSI)          19

The following lemmas will be used to make some theoretical comparisons among estimators in the next section.

Lemma 1: Let n x n matrices M > 0, N > 0 (or N > 0), then N > M if and only if λ 1 N M 1  < 1 where λ 1 N M 1  < 1 is the largest eigenvalue of matrix NM-1 [22].

Lemma 2: Let M be an n x n positive definite matrix, that is M > 0, and α be some vector, then Mαα'0 if and only if α' M 1 α1 [23].

Lemma 3: Let α ^ i  =  A i y i = 1, 2 be two linear estimators of α. Suppose that D = cov α ^ 1 cov α 2  > 0, where cov α ^ 1 i = 1, 2 denotes the covariance matrix of α ^ i and b i  = Bias α ^ i =( A i XI)α, i = 1, 2. Consequently,

Δ α ^ 1 α ^ 2  = MSEM α ^ 1 MSEM α ^ 2 = σ 2 D+ b 1 b 1 ' b 2 b 2 ' >0         20

If and only if b 2 ' σ 2 D+ b 1 b 1 ' 1 b 2 <1, where MSEM α ^ i =cov α ^ i + b i b i ' [24].

Comparison among the Estimators

Theoretical Comparisons among the Proposed LDK Estimator and the OLSE, Ordinary Ridge Regression (ORR), Liu Estimators, Kibria-Lukman (KL) Estimator, Dawoud-Kibria (DK), and Modified Ridge-Type (MRT) Estimator.

Theorem 1: The estimator α ^ LDK is superior to the estimator α ^ if and only if α' WSI ' σ 2 Λ 1 WS Λ 1 WS α WSI  < 1      21

Proof: The difference between the dispersion matrices is given as D α ^ D α ^ LDK = σ 2 Λ 1 WS Λ 1 WS

= σ 2 diag 1 λ i λ i +d 2 λ i k 1+d 2 λ i λ i +1 2 λ i +k 1+d 2 i=1 p            22

Λ 1 WS Λ 1 WS will be positively definite if and only if λ i +1 2 λ i +k 1+d 2 λ i +d 2 λ i k 1+d 2 >0.

Theorem 2: The estimator α ^ LDK is superior to the estimator α ^ k if and only if α' WSI ' σ 2 AΛAWS Λ 1 WS α WSI  < 1     23

Proof: The difference between the dispersion matrices is given as

D α ^ k D α ^ LDK = σ 2 AΛAWS Λ 1 WS

= σ 2 diag λ i λ i +k 2 λ i +d 2 λ i k 1+d 2 λ i λ i +1 2 λ i +k 1+d 2 i=1 p              24

AΛA - WSΛ-1WS will be positively definite if and only if

λ i 2 λ i +1 2 λ i +k 1+d 2 λ i +d 2 λ i +k 2 λ i k 1+d 2 >0.

Theorem 3: The estimator α ^ LDK is superior to the estimator α ^ d if and only if α' WSI ' σ 2 E Λ 1 EWS Λ 1 WS α WSI  < 1           25

Proof: The difference between the dispersion matrices is given as D α ^ d D α ^ LDK = σ 2 E Λ 1 EWS Λ 1 WS

= σ 2 diag λ i +d 2 λ i λ i +1 2 λ i +d 2 λ i k 1+d 2 λ i λ i +1 2 λ i +k 1+d 2 i=1 p         26

EΛ-1E - WSΛ-1WS will be positively definite if and only if

λ i +d 2 λ i +k 1+d 2 λ i +d 2 λ i k 1+d 2 >0.

Theorem 4: The estimator α ^ LDK is superior to the estimator α ^ KL if and only if α' WSI ' σ 2 P Λ 1 PWS Λ 1 WS α WSI  < 1       27

Proof: The difference between the dispersion matrices is given as D α ^ KL D α ^ LDK = σ 2 P Λ 1 PWS Λ 1 WS

= σ 2 diag λ i k 2 λ i λ i +k 2 λ i +d 2 λ i k 1+d 2 λ i λ i +1 2 λ i +k 1+d 2 i=1 p           28

PΛ-1P - WSΛ-1WS will be positively definite if and only if

λ i k 2 λ i +1 2 λ i +k 1+d 2 λ i +d 2 λ i +k 2 λ i k 1+d 2 >0.

Theorem 5: The estimator α ^ LDK is superior to the estimator α ^ MRT if and only if α' WSI ' σ 2 H Λ 1 HWS Λ 1 WS α WSI  < 1        29

Proof: The difference between the dispersion matrices is given as

D α ^ MRT D α ^ LDK = σ 2 H Λ 1 HWS Λ 1 WS

= σ 2 diag λ i λ i +k 1+d 2 λ i +d 2 λ i k 1+d 2 λ i λ i +1 2 λ i +k 1+d 2 i=1 p            30

HΛ-1H - WSΛ-1WS will be positively definite if and only if λ i 2 λ i +1 2 λ i +d 2 λ i k 1+d 2 >0.

Theorem 6: The estimator α ^ LDK is superior to the estimator α ^ DK if and only if

α' WSI ' σ 2 P Λ 1 PWS Λ 1 WS α WSI  < 1         31

Proof: The difference between the dispersion matrices is given as D α ^ DK D α ^ LDK = σ 2 KM Λ 1 KMWS Λ 1 WS

= σ 2 diag λ i k 1+d 2 λ i λ i +k 1+d 2 λ i +d 2 λ i k 1+d 2 λ i λ i +1 2 λ i +k 1+d 2 i=1 p           32

KMΛ-1KM - WSΛ-1WS will be positively definite if and only if

λ i +1 2 λ i k 1+d 2 λ i +d 2 λ i k 1+d 2 >0.

Determination of the Parameters k and d

Various researchers have introduced different estimators of k and d for different kinds of regression models. Some of these authors are Hoerl and Kennard [4]; Dorugade [25]; Lukman and Ayinde [26]; Aslam and Ahmad [27], among others.The optimal values of k and d for the proposed estimator are obtainedthus. In determining the optimal value of k, d is fixed. The optimal value of the k can be considered to be that k that minimizes:

MSEM α ^ LDK  =  σ 2 WS Λ 1 WS + (WSI)αα'(WSI)

g k,d =MSEM α ^ LDK =tr MSEM α ^ LDK

g k,d = σ 2 i p λ i +d 2 λ i k 1+d 2 λ i λ i +k 1+d 2 λ i +1 2 + i p 2k λ i + λ i d+d +k d 2 +1 + λ i 1d 2 α i 2 λ i +k 1+d 2 λ i +1 2        33

Taking the partial derivative of the function g(k,d) with respect to k gives g k,d k =2 σ 2 i=1 p λ i +d 2 λ i k d+1 d+1 λ i λ i +k d+1 2 λ i +1 2 2 σ 2 i=1 p λ i +d 2 λ i k d+1 2 d+1 λ i λ i +k d+1 3 λ i +1 2 +2 i=1 p 2k λ i + λ i d+d λ i d1 +k d 2 +1 α i 2 λ i +k d+1 2 λ i +1 2 2 i=1 p 2k λ i + λ i d+d λ i d1 +k d 2 +1 2 d+1 α i 2 λ i +k d+1 3 λ i +1 2    34

Let g k,d k =0;

k= σ 2 λ i λ i +d α i 2 λ i 2 1d σ 2 d d+1 + σ 2 λ i d+1 + α i 2 λ i d 2 +1 +2 α i 2 λ i λ i d+d+ λ i       35

For practical purposes, σ2 and α i 2 are replaced with σ ^ 2 and α ^ i 2 , respectively. Consequently, (30) becomes

k ^ = σ ^ 2 λ i λ i +d α ^ i 2 λ i 2 1d σ ^ 2 d d+1 + σ ^ 2 λ i d+1 + α ^ i 2 λ i d 2 +1 +2 α ^ i 2 λ i λ i d+d+ λ i      36

and,

k ^ min =min σ ^ 2 λ i λ i +d α ^ i 2 λ i 2 1d σ ^ 2 d d+1 + σ ^ 2 λ i d+1 + α ^ i 2 λ i d 2 +1 +2 α ^ i 2 λ i λ i d+d+ λ i i=1 p            37

and the biasing parameters d proposed by Liu [5] in equation (38) will be adopted in this study. It is given as follows:

d= λ i α i 2 σ 2 σ 2 + λ i α i 2      38

For practical purposes, σ 2 and α i 2 can be considered as σ ^ 2 and α ^ i 2 , respectively. Consequently, (38) becomes

d= λ i α ^ i 2 σ ^ 2 σ ^ 2 + λ i α ^ i 2         39

Simulation and Application

Simulation technique

The simulation procedure used by McDonald and Galarneau [28]; Wichern and Churchill [29]; Gibbons [30]; Lukman and Ayinde [26] was utilized to generate the predictor variables in this study: This is given as:

x ij = 1 ρ 2 1 2 z ij +ρ z i,p+1 , i=1,2,,n, j=1,2,,p.      40

where zij is an independent standard normal distribution with mean zero and unit variance, ρ is the correlation between any two explanatory variables, and p is the number of explanatory variables. For this study, we considered the values of ρ to be 0.8, 0.9, 0.95, and 0.99. Also, explanatory variables (p) were taken to be three (3) and seven (7) for the simulation study. The error terms ut were generated following Firinguetti [31] such that u t ~N 0, σ 2 I . The values of β'β=1 Newhouse and Oman [32]. The standard deviations in this simulation study were σ = 3, 5, and 10.

Simulation results from discussion

Table 2 and Table 3 show that as σ and ρ increase, the estimated MSE values increase. Likewise, as n increases, the estimated MSE values decrease. As expected and from several simulations and empirical research, when the multicollinearity problem exists, the OLS estimator gives the highest MSE values and performs the worst among all estimators. Additionally, the results show that the proposed LDK estimator performs best and dominates the rest of the estimators in this study. Thus, the findings agree with the theoretical results.

Table 2: Estimated MSE when p = 3, n = 50. View Table 2

Table 3: Estimated MSE when p = 3, n = 100. View Table 3

Numerical example

In this section, Portland cement data was used to demonstrate the performance of the proposed estimator. The Portland cement data was originally adopted by Woods, et al. [33] and was later adopted by Li and Yang [34] and Ayinde, et al. [35]. The data set is widely known as the Portland cement dataset. The regression model for these data is defined as:

y i = β 1 X 1 + β 2 X 2 + β 3 X 3 + β 4 X 4 + ε i      41

where yi = heat evolved after 180 days of curing measured in calories per gram of cement, X1 = tricalcium aluminate, X2 = tricalcium silicate, X3 = tetracalcium aluminoferrite, and X4 = β-dicalcium silicate. The variance inflation factors (VIFs) are 38.50, 254.42, 46.87, and 282.51, respectively. Eigenvalues of X'X matrix are λ1 = 44676.206, λ2 = 5965.422, λ3 = 809.952, and λ4 = 105.419, and the condition number of X'X is approximately 424. The VIFs, the eigenvalues, and the condition number indicate severe Multicollinearity. The estimated parameters and the MSE values of the estimators are presented in Table 4.

From Table 4, the proposed estimator (LDK) performs best among other estimators as it gives the smallest MSE value. Just as observed in the simulation study results, the OLS estimator did not perform well in the presence of Multicollinearity as it has the highest MSE.

Table 4: The results of regression coefficients and the corresponding MSE values. View Table 4

Concluding Remarks

In this paper, a new two-parameter estimator (LKD) is proposed. Theoretical comparison of the proposed with six other existing estimators shows the superiority of the proposed estimator. Results from the simulation study reveal that the proposed estimator performs better than other existing estimators used in this study under certain conditions, which further strengthens the theoretical study. Application to real-life dataset also reveals the dominance of the proposed estimator (LKD). The newly proposed estimator (LKD) is recommended for parameter estimation in the linear regression model in the presence of Multicollinearity.

Declaration of Competing Interest

The authors declare that they have no known competing interests.

Funding

No financial assistance was received for this study.

References

  1. Frisch R (1934) Statistical confluence analysis by means of complete regression analysis. University Institute of Economics, Oslo.
  2. Belsley DA, Kuh E, Welsch RE (1980) Regression Diagnostics: Identifying influential data and sources of collinearity. John Wiley & Sons, New York.
  3. Hawking RR, Pendleton OJ (1983) The regression dilemma. Communication in Statistics Theory and Method 12: 497-527.
  4. Hoerl AE, Kennard RW (1970) Ridge regression: Biased estimation for nonorthogonal problems. Technometrics 12: 55-67.
  5. Liu K (1993) A new class of biased estimate in linear regression. Communications in Statistics - Theory and Methods 22: 393-402.
  6. Kibria BM (2003) Performance of some new ridge regression estimators. Communications in Statistics -Simulation and Computation 32: 419-435.
  7. Sakallioglu S, Kaciranlar S (2008) A new biased estimator based on ridge estimation. Statistical Papers 49: 669-689.
  8. Baye MR, Parker DF (1984) Combining ridge and principal component regression: A money demand illustration. Communications in Statistics - Theory and Methods 13: 197-205.
  9. Yang H, Chang X (2010) A new two-parameter estimator in linear regression. Communications in Statistics - Theory and Methods 39: 923-934.
  10. Wu J, Yang H (2011) Efficiency of an almost unbiased two-parameter estimator in linear regression model. Statistics 47: 535-545.
  11. Dorugade A (2014) A modified two-parameter estimator in linear regression. Statistics in Transition 15: 23-36.
  12. Lukman AF, Ayinde K, Binuomote S, Clement OA (2019a) Modified ridge-type estimator to combat multicollinearity: Application to chemical data. Journal of Chemometrics 33: e3125.
  13. Lukman AF, Ayinde K, Sek SK, Adewuyi E (2019b) A modified new two-parameter estimator in a linear regression model. Modelling and Simulation in Engineering 2019: 1-10.
  14. Kibria BMG, Lukman AF (2020) A new ridge-type estimator for the linear regression model: Simulations and applications. Scientifica 2020: 1-16.
  15. Dawoud I, Kibria BMG (2020) A new biased estimator to combat the multicollinearity of the gaussian linear regression model. Stats 3: 526-541.
  16. Owolabi AT, Ayinde K, Alabi OO (2022) A new ridge-type estimator for the linear regression model with correlated regressors. Concurrency Computat Pract Exper 34: e6933.
  17. Owolabi AT, Ayinde K, Idowu JI, Oladapo OJ, Lukman AF (2022) A new two-parameter estimator in the linear regression model with correlated regressors. J Stat Appl Pro 11: 499-512.
  18. Xin Y, Xiao-Gang S (2009) Linear regression analysis: Theory and computing. World Scientific Publishing Co. Pte. Ltd.
  19. Chatterjee S, Hadi AS (2006) Regression analysis by example. (4th edn), John Wiley and Sons, NJ.
  20. Kaciranlar S, Sakallioglu S, Akdeniz F, Styan GPH, Werner HJ (1999) A new biased estimator in linear regression and a detailed analysis of the widely-analysed dataset on Portland Cement. Sankhya: Indian Journal of Statistics 61: 443-459.
  21. Dawoud I, Lukman AF, Abdul-Rahaman H (2022) A new biased regression estimator: Theory, simulation and application. Scientific African 15: e01100.
  22. Wang SG, Wu MX, Jia ZZ (2006) Matrix inequalities. (2nd edn), Chinese Science Press, Beijing.
  23. Farebrother RW (1976) Further results on the mean square error of ridge regression. J R Stat Soc 38: 248‐250.
  24. Trenkler G, Toutenburg H (1990) Mean squared error matrix comparisons between biased estimators-an overview of recent results. Stat Pap 31: 165‐179.
  25. Dorugade AV, Kashid DN (2010) Alternative method for choosing ridge parameter for regression. International Journal of Applied Mathematical Sciences 4: 447-456.
  26. Lukman AF, Ayinde K (2017) Review and classifications of the ridge parameter estimation techniques. Hacettepe Journal of Mathematics and Statistics 46: 953‐967.
  27. Ahmad S, Aslam M (2020) Another proposal about the new two-parameter estimator for linear regression model with correlated regressors. Communications in Statistics - Simulation and Computation 51: 3054-3072.
  28. McDonald GC, Galarneau DI (1975) A Monte Carlo evaluation of some ridge-type estimators. Journal of the American Statistical Association 70: 407-416.
  29. Wichern DW, Churchill GA (1978) A comparison of ridge estimators. Technometrics 20: 301-311.
  30. Gibbons DG (1981) A simulation study of some ridge estimators. Journal of the American Statistical Association 76: 131-139.
  31. Firinguetti LL (1989) A simulation study of ridge regression estimators with autocorrelated errors. Communications in Statistics Simulation and Computation 18: 673-702.
  32. Newhouse JP, Oman SD (1971) An evaluation of ridge estimators. A report prepared for United States Air Force Project RAND.
  33. Woods H, Steinour HH, Starke HR (1932) Effect of composition of Portland cement on heat evolved during hardening. Ind Eng Chem 24: 1207-1214.
  34. Li Y, Yang H (2012) A new Liu-type estimator in linear regression model. Stat 53: 427-437.
  35. Ayinde K, Lukman AF, Samuel OO, Ajiboye S (2018) Some new adjusted ridge estimators of linear regression model. Int J Civ Eng Technol 9: 2838‐2852.

Citation

Oladapo OJ, Owolabi AT, Idowu JI, Ayinde K (2022) A New Modified Liu Ridge-Type Estimator for the Linear Regression Model: Simulation and Application. Int J Clin Biostat Biom 8:048. doi.org/10.23937/2469-5831/1510048