International Journal of Statistics and Applications
p-ISSN: 2168-5193 e-ISSN: 2168-5215
2016; 6(2): 35-39
doi:10.5923/j.statistics.20160602.01

Ismail B. 1, Manjula Suvarna 2
1Department of Statistics, Mangalore University, Mangalagangothri, Mangalore, India
2Department of Community Medicine, A.J. Institute of Medical Sciences and Research Centre, Mangalore, India
Correspondence to: Manjula Suvarna , Department of Community Medicine, A.J. Institute of Medical Sciences and Research Centre, Mangalore, India.
| Email: | ![]() |
Copyright © 2016 Scientific & Academic Publishing. All Rights Reserved.
This work is licensed under the Creative Commons Attribution International License (CC BY).
http://creativecommons.org/licenses/by/4.0/

When using the linear statistical model, researchers face variety of problems due to non experimental nature i.e uncertainity about the nature of the error process, model mis- specifications, dependent regressors etc. The phenomenon of correlated errors in linear regression models involving time series data is called autocorrelation. Violation of the assumption of independent regressors leads to multicollinearity. Hence, Ordinary ridge estimates are imprecise to be of much use in case of autocorrelated regression model with the multicollinearity problem. Objective: To develop a new estimator for the regression parameter in the presence of multicollinearity and autocorrelation. To choose an appropriate ridge parameter for the proposed estimator using Monte Carlo simulation. Materials and Methods: Monte Carlo simulation study is carried out using the Statistical programming language MATLAB version 7.0 to evaluate the performance of the proposed estimator based on the Mean squared error (MSE) criterion. Findings: Determined the regions where a particular method for estimating ridge parameter performs better among different existing methods. This estimate of ridge parameter is used in the proposed estimator. The proposed estimator performs better than the existing estimator under the MSE criterion.
Keywords: Correlated regressors, Autocorrelation, Mean Squared error, Monte Carlo Simulation
Cite this paper: Ismail B. , Manjula Suvarna , Estimation of Linear Regression Model with Correlated Regressors in the Presence of Autocorrelation, International Journal of Statistics and Applications, Vol. 6 No. 2, 2016, pp. 35-39. doi: 10.5923/j.statistics.20160602.01.
![]() | (1) |
with the covariance matrix of
is obtained as
.It is not necessary that assumptions mentioned above hold good in real life situation. The regressors may be nearly correlated and the responses may also be correlated. In such instances the OLS estimator mentioned above do not possess the optimum statistical property. Hence there is a need to develop a new estimator which takes care of this situation. Ridge Regression: The violation of the assumption of independent regressors leads to multicollinearity. If X is less than full rank then such a situation is known as perfect multicollinearity. In this case OLS estimator does not exist. This situation is very rare in practice. In most of the real life situations, some regressors are nearly related to the remainining regressors. This is known as near multicollinearity. In case of near multicollinearity, rank of the regressor matrix X is equal to k and hence OLS estimator exist, but they are too imprecise to be of much use [2]. With strongly interrelated pairs of regressors, X’X is illconditioned and the variance of the OLS estimator becomes large. With multicollinearity, the estimated OLS coefficients may be statistically insignificant (too large, too small and even have wrong sign). Hence interpretation given to the regression coefficients may no longer be valid. It may be preferable to consider biased estimators of β, if their variances are sufficiently smaller than those of OLS estimators. One such biased estimator is the “ridge estimator”. The ridge estimator (ordinary ridge estimator) of β is![]() | (2) |
, where
represents the error variance of model (1),
is the maximum among elements of
defined as
=D’β with D being an orthogonal matrix. Autocorrelation: Autocorrelation is said to exist when the successive observations in linear regression model are correlated. The existence of autocorrelated errors has been rationalized in a variety of ways, as noted by Maddala [9].The successive dependence of the error term is represented by![]() | (3) |
[10]. When the error satisfies the relation (3), the observations follow first order autocorrelation. The variance covariance matrix of Y is
Where
and
Since the covariance matrix of ε is nonspherical (i.e not a scalar multiple of the identity matrix), OLS, though unbiased, is inefficient relative to generalised least squares by Aitken’s theorem. The generalized least squares estimator of β in (1) is [10]![]() | (4) |
in (3) is known then we can write ![]() | (5) |
. Hence autocorrelation is a particular case of heteroscedasticity. In the case of heteroscedasticity, GLS is an appropriate method of estimation as given in (4). Further, when there is multicollinearity, often used method is the ridge regression as mentioned in (2). Combining these two methods, we propose for the autocorrelated model with multicollinearity a generalized ridge type estimator represented as 
where
is as defined in (5).Hence the model under consideration contains the unknown parameters k,
,
and β.In the following [11] we present some existing methods for estimating ridge parameter k 1. Hoerl and Kennard method ![]() | (6) |
![]() | (7) |
![]() | (8) |
are the eigen values of
.
X=[ X0 , X1 , X2 , X3 ], β=[ β0 β1 β2 β3 ]’ =[4 2.5 1.8 0.6]’, X0 =[1 1 1….1]’.To generate normally distributed random variables X1, X2, X3 with specified intercorrelations we use the following equations [12, 13].
where 
In equation (3), ut are independent and identically distributed normal random variables with mean 0 and variance
. The autocorrelation coefficient
in (3) is ranging from -0.9 to -0.1 and the regression parameters are fixed as
,
. The parameters of the model in equation (4) are fixed as
. Taking
sixteen different levels of intercorrelation (multicollinearity) among the regressors are taken as -0.2, -0.3, -0.4, -0.5, -0.6, -0.7, -0.8, -0.9, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9. With the above setup a sample of 100 observations are generated and replicated 1000 times. For each choice of the k, the MSE for the generalized ridge estimator is computed. The estimator of the ridge parameter (k) which gives minimum MSE is recorded for different choice of the parameters
and the results are presented in Table 1.
|
| [1] | Gunst, R..F and Manson, R. L, 1979, Some considerations in the evaluation of alternate prediction equations, Technometrics, 21: 55-63. |
| [2] | Peter Schmidt, 1976,. Econometrics, New York: Marcel dekker Inc. |
| [3] | Hoerl, A. E. and Kennard, R.W, 1970, Ridge regression: biased estimation for nonorthogonal problems, Technometrics, 12: 9-82. |
| [4] | Hoerl, A.E., Kennard, R. W. and Baldwin, K. F., 1975, Ridge regression: Some simulation. Communications in Statistics, 4: 105-123. |
| [5] | Mc Donald, G.C and Galarneau, D. I, 1975, A Monte carlo evaluation of some ridge type estimators. Journal of American statistical Association, 70(350): 407-412. |
| [6] | Hocking, R.R., F.M. Speed and M.J. Lynn, 1976, A class of biased estimators in linear regression. Technometrics, 18: 425-437. |
| [7] | Saleh, A.K. and Kibria, B. M., 1993, Performances of some new preliminary test ridge regression estimators and their properties. Communications in Statistics- Theory and Methods, 22: 2747-2764. |
| [8] | Khalaf, G. and Shukur, G., 2005, Choosing ridge parameter for regression problems. Communications in Statistics- Theory and Methods, 34: 1177-1182. |
| [9] | Maddala, G, 1977, Econometrics, New York: McGraw-Hill. |
| [10] | Thomas B. Fomby, R. Carter Hill, Stanley R. Johnson, 1984, Advanced Econometric Methods, New York:Springer Verlag. |
| [11] | Al-Hassan, Yazid M., 2010, Performance of a new Ridge Regression Estimator. Journal of Association of Arab Universities for Basic and Applied Sciences, 9:43-50. |
| [12] | K. Ayinde and O. S. Adegboye, 2010, Equations for generating normally distributed random variables with specified intercorrelation. Journal of Mathematical Sciences, 21: 83-203. |
| [13] | Kayode Ayinde, Emmanuel O. Apata, Oluwayemisi O. Alaba, 2012, Estimators of Linear Regression model and prediction under some assumptions violation. Journal of Statistics, 2:534-546. |