Polish Statistical Association
Central Statistical Office of Poland
Subject: Economics, Statistics & Probability
ISSN: 1234-7655
eISSN: 2450-0291
SEARCH WITHIN CONTENT
Shalini Chandra * / Gargi Tyagi *
Keywords : omission of relevant variables, multicollinearity, r − (k, d) class estimator, mean squared error
Citation Information : Statistics in Transition New Series. Volume 18, Issue 1, Pages 27-52, DOI: https://doi.org/10.21307/stattrans-2016-056
License : (CC BY 4.0)
Published Online: 17-July-2017
In this paper, the effect of misspecification due to omission of relevant variables on the dominance of the
In multiple linear regression, the presence of multicollinearity inflates sampling variance of the ordinary least squares estimator and may also produce wrong signs of the estimator. Many authors have witnessed the presence of multicollinearity in the various fields of application, including Hamilton (1972), Mahajan et al. (1977), Heikkila (1988), Graham (2003), among others. To cope up with the problem of multicollinearity several alternative methods to the OLS have been proposed, viz. ordinary ridge regression (ORR) by Hoerl and Kennard (1970); principal component regression (PCR) by Massy (1965). In the hope that combining two estimators will contain the properties of both gave rise to the development of the r − k class, the two-parameter class and the r − (k, d) class estimators (see Baye and Parker (1984); Kaciranlar and Sakallioglu (2001); Özkale and Kaciranlar (2007) and Özkale (2012)).
The performance of these estimators have been evaluated under various comparison criteria like mean squared error (MSE), matrix MSE, Pitman’s closeness criterion and the Mahalanobis loss function. Nomura and Ohkubo (1985) derived the dominance conditions of the r − k class estimator over the OLS and ORR estimators and Sarkar (1996) obtained conditions of the superiority of the r − k class estimator over the other estimators under matrix MSE criterion. Özkale and Kaciranlar (2008) compared the r − k class estimator with the OLS estimator under Pitman’s closeness criterion. Özkale (2012) proposed the r − (k, d) class estimator and compared this estimator with the other biased estimators under the MSE criterion. Sarkar and Chandra (2015) studied the performance of the r − (k, d) class estimator over the OLS, PCR and the two-parameter class estimator under the Mahalanobis loss function and derived tests to verify the conditions.
In these studies, it has been assumed inherently that the model is correctly specified. However, in practice, some of the relevant regressors may get excluded from the model, i.e. the model does not remain correctly specified, known as misspecified model. The omission of relevant regressors causes biased and inconsistent estimation. The effect of the omission of relevant regressors on the performance of the estimators have been studied by several authors, for example, Kadiyala (1986); Trenkler and Wijekoon (1989) and Wijekoon and Trenkler (1989). Although not much work has been done when some of the regressors are omitted and multicollinearity is also present, Sarkar (1989) studied the performance of the r − k class estimator and compared it with the OLS, ORR and PCR estimators under MSE criterion when the model is misspecified due to omission of relevant regressors.
In this paper, misspecification due to omission of relevant regressors and multicollinearity have been studied simultaneously and the effect of misspecification on the dominance of the r − (k, d) class estimator over the other biased estimators has been studied under the MSE criterion. The plan of this paper is as follows: in Section 2, the model and the estimators under study are given. Section 3 provides the comparison of the estimators and a Monte Carlo simulation has been given in Section 4. A numerical example is given in Section 5 to see the effect of misspecification on the estimators, which in turn exhibits the utility of the estimators. The paper is concluded in Section 6.
Let us consider the regression model as:
where y is an n × 1 vector of dependent variable, X and Z are n × p and n × q full column rank matrices of regressors respectively such that X′X and Z′Z are ill-conditioned, p + q < n, β and γ are the corresponding p × 1 and q × 1 vectors of parameters associated with X and Z, respectively. ε is an n × 1 vector of disturbance term, and it is assumed that ε~N(0, σ2In). Suppose that an investigator has unknowingly excluded regressors of Z matrix, thus the misspecified model is given by: where u = Zγ + ε. Misspecification occurs when the investigator assumes the disturbance vector u to be normally distributed with mean vector 0 and variance σ2In.Let us consider the following transformation for the model in (2.2):
where X* = XT, T′β = α, T = (t1, t2, …, tp) is a p × p orthogonal matrix with T′X′XT = Λ and Λ = diag(λ1, λ2, …, λp) is a p × p diagonal matrix of eigen values of X′X matrix such that λ1, ≥ λ2, ≥ … ≥ λp. Now, let Tr = (t1, t2, …, tr) be p × r orthogonal matrix after deleting last p − r columns from T matrix, where r ≤ p. Thus, Tr′X′XTr = Λr where Λr = diag(λ1, λ2, …, λr) and Tp−r′X′XTp−r = Λp−r, where Λp−r = diag(λr+1, λr+2, …, λp). Also, T′T = Tr′Tr + Tp−r′Tp−r and let N = {1, 2, …, r; r + 1, …, p} be a set of first p integers such that N = {Nr; Np−r} where Nr = {1, 2, …, r} and Np−r = {r + 1, r + 2, …, p}.Özkale (2012) introduced an estimator by grafting the two-parameter class estimator and the PCR estimator together, known as the r − (k, d) class estimator to deal with the problem of multicollinearity. For the misspecified model in (2.2) the r − (k, d) class estimator is given by:
which can be rewritten as: where Sr(k) = Λr + kIr and Sr(kd) = Λr + kdIr. This is a general estimator which includes the OLS, ORR, PCR, r − k class and the two-parameter class estimators as its special cases as:, is the OLS estimator,
, is the ORR estimator,
, is the PCR estimator,
, is the r − k class estimator,
, is the two-parameter class estimator.
From (2.5), the bias and the variance of can be obtained as:
where δ = X′Zγ, and respectively.It is clear from (2.6) and (2.7) that the bias of the r − (k, d) class estimator increases due to omission of relevant regressors whereas the variance of the estimator is not affected by the misspecification.
Further, the MSE for an estimator of β is defined as:
By substituting (2.6) and (2.7) in (2.8) and on simplification, we get:
which can be rewritten as: where T′δ = η = {η1, η2, …, ηp}. Following Özkale(2012), the first under-bracket is the MSE obtained when there is no misspecification and the second under-bracket is the contribution of omission of relevant regressors.The MSE of other estimators can be obtained by substituting the suitable values of r, k and d in (2.10). From the risk expression in (2.10), it can be seen that the effect of omission of relevant regressors on the MSE values will depend on the sign of the second term. If αiηi is negative for all values of i ∈ Nr, the second term in (2.10) will be negative and thus the MSE of the r − (k, d) class estimator will increase due to omission of relevant regressors. However, if αiηi is non-negative for some values of i ∈ Nr no definite conclusion can be made regarding the effect of misspecification.
The selection of the unknown biasing parameters k and d in the r − (k, d) class estimator is an important problem. The optimum values of k and d in r − (k, d) class estimator can be obtained by minimizing the MSE of the estimator with respect to k and d. To find a pair (k, d) of optimum values of k and d, we will use the technique of maxima and minima in calculus.
Let the two-dimensional function MSE((k, d)) have its minimum value at (k0, d0) and have a continuous partial derivative at this point, then and . The points k0 and d0 can be found as follows:
On differentiating MSE((k, d)) in (2.10) with respect to d keeping r and k fixed, we obtain
and equating (2.11) to zero, we get:Assuming that αiηi > 0 for all i ∈ Nr, if and λi > k for all i ∈ Nr, the upper bound of d0 is given by
which is the optimum value of d when there is no misspecification due to omission of relevant regressors. Thus, if kαi2 − σ2 < 0 and λi > k for all i ∈ Nr, the optimum value of d in the misspecified model is less than that in the case of no misspecification. Moreover, for d0 to be a positive value should be positive for i ∈ Nr.Further, differentiating (2.10) with respect to k keeping r and d fixed, we obtain:
From (2.14) and (2.12), we have:
Clearly, is zero, when
Then (k0, d0) is the expected point which minimizes MSE((k, d)) where k0 and d0 are given as (2.16) and (2.12) respectively. However, when we substitute k = k0 in (2.12) d0 becomes zero. Therefore, a point (k0, d0) which satisfies k < 0, 0 < d < 1 and minimizes MSE((k, d)) cannot be found (see Fig.1). In order to find an appropriate value of k and d, the behaviour of the MSE of the estimator at boundary points can be studied. This conclusion has been illustrated through the graph reported below:
From Figure 1 the effect of misspecification on the optimum values of k and d for fixed values of d and k respectively can be observed, and, also the pair of values of k and d may not be found out for which the r − (k, d) class estimator has minimum MSE. Further, we note that for the fixed values of d, the MSE of takes the minimum value for smaller value of k in the misspecified model when compared with the true model. However, for small value of k (see Fig. (a) and Fig. (b)), no variations are observed in the MSE values of for both the models, whereas for k = 5, the MSE of takes the minimum value for a smaller value of d in the misspecified model.
In this section, we compare the r − (k, d) class estimator with other biased estimators when the model is misspecified due to omission of relevant regressors, and also study the effect of misspecification on the dominance conditions.
The MSE of the OLS estimator in the misspecified model can be obtained by substituting r = p, k = 0 in (2.10), as:
The difference of MSEs of the r − (k, d) class estimator and the OLS estimator, say Δ1, can be written as:
On further simplification, the difference can be rewritten as:
It is clear from the above expression that Δ1 ≥ 0 that is, the r − (k, d) class estimator dominates the OLS estimator, for all k < 0, 0 < d < 1 if , for all i ∈ N. From (3.3), it can also be observed that when there is no misspecification due to omission of relevant regressors (i.e. ηi = 0 for all i ∈ N) the condition reduces to for all i ∈ N, which is the same as that of obtained by Ozkale (2012)). It is evident that due to addition of a positive term the odds for Δ1 ≥ 0 are higher in the misspecified model.
Further, if , i ∈ Nr and , ∀ i ∈ Np−r then for a fixed k there exists a d in the range:
such that the r − (k, d) class estimator dominates the OLS estimator. If αiηi < 0 for all i ∈ Nr, then the lower limit of d decreases due to omission of relevant regressors and thus the dominance range of the r − (k, d) class estimator over the OLS estimator increases.Furthermore, if , for some i = r + 1, r + 2, …, p, no definite conclusion can be drawn regarding dominance of one over the other. The results obtained are reported in the form of the following theorem.
(i) If , for all i ∈ N, the r − (k, d) class estimator dominates the OLS estimator for all k > 0 and 0 < d < 1. The odds for superiority of r − (k, d) class estimator over the OLS estimator increases in the misspecified model.
(ii) If , i ∈ Nr and , for all i ∈ Np−r, the r − (k, d) class estimator dominates the OLS estimator for all k > 0 and d such that it satisfies (3.4). The range of dominance of r − (k, d) class estimator increases in the misspecified model provided αiηi is positive for all i ∈ Nr.
(iii) If , for some i = r + 1, r + 2, …, p, no definite conclusion can be drawn regarding their dominance.
The MSE of the ORR estimator can be obtained by substituting r = p and d = 0 in (2.10), given as:
Using (2.10) and (3.5), the difference between the MSEs,; say Δ2, is given by:
On further simplification, we obtain:
From (3.7), it can be noticed that the r − (k, d) class estimator dominates the ORR estimator if both summations are positive, that is:
andIf for all i ∈ Np−r then (3.9) holds when:
If for all i ∈ Np−r, then (3.9) does not hold true. However, if for some i ∈ Np−r, a positive k can be found such that the second summation in Δ2, i.e. is positive.
Further, from (3.8) we obtain:
For d to be a positive in (3.11), should be a positive value for all i ∈ Nr, i.e.
If upper bound of d in (3.11) is greater than 1, any value smaller than 1 can be taken, which satisfies (3.11) and 0 < d < 1.
The conditions of dominance of the r − (k, d) class estimator over the ORR estimator under MSE criterion is stated below in the form of the following theorem:
(i) If for all i ∈ Nr and for all i ∈ Nr, the r − (k, d) class estimator dominates the ORR estimator if and 0 < d < .
(ii) If for some i ∈ Np−r and for all i ∈ Nr, the r − (k, d) class estimator dominates the ORR estimator for a value of k such that is positive and .
On substituting k = 0 in (2.10), the MSE of the PCR estimator can be obtained as:
From (2.10) and (3.13), the difference in MSEs, say Δ3, is given by:
On further simplifying it, we get:
It can be observed from the above expression that if for all i ∈ Nr, the r − (k, d) class estimator dominates the PCR estimator for all k > 0 and 0 < d < 1. Evidently, the odds for superiority of the r − (k, d) class estimator over the PCR estimator increases due to misspecification.
If for all i ∈ Nr, then Δ3 is positive when is positive for all i ∈ Nr, which can be rewritten as
If , forall i ∈ Nr, i.e.
then Δ3 is positive for all k < 0. It is noticeable that the lower limit of d decreases due to misspecification, thus a wider range for the dominance of the r − (k, d) class estimator over the PCR estimator is obtained as compared with no misspecification.Further, if
holds, then Δ3 is positive for k such thatBy rewriting (3.18), it is observed that the upper limit of k increases due to misspecification. Additionally, the upper limit of d decreases. Thus, due to misspecification a wider range of k for a shorter range of d in which the r − (k, d) class estimator dominates the PCR estimator is obtained.
The comparisons can be concluded in the following theorem.
(i) If for all i ∈ Nr, the r − (k, d) class estimator dominates the PCR estimator for all k > 0 and 0 < d < 1. The odds for superiority of the r − (k, d) class estimator over the PCR estimator increases due to misspecification.
(ii) If for all i ∈ Nr and , then the r − (k, d) class estimator dominates the PCR estimator for all k > 0. The dominance range of the r − (k, d) class estimator over the PCR estimator increases due to misspecification.
(iii) If for all i ∈ Nr and , then the r − (k, d) class estimator dominates the PCR estimator if . The range of k increases while the range of d decreases due to misspecification, in which the r − (k, d) class estimator dominates the PCR estimator.
The MSE of the r − k class estimator can be obtained by substituting d = 0 in (2.10), given as:
From (2.10) and (3.19), the difference between the MSEs, say Δ4, is given by:
On further simplification, we get:
From (3.21), the r − (k, d) class estimator dominates the r − k class estimator when
If for all i ∈ Nr and for all i ∈ Nr, that is:
then Δ4 is positive when k is such that:If for all i ∈ Nr and for all i ∈ Nr, that is:
then Δ4 is negative for all k < 0.If for all i ∈ Nr, then Δ4 is negative for all k < 0 and 0 < d < 1.
The results obtained are given in the following theorem.
(i) If for all i ∈ Nr and , then the r − (k, d) class estimator dominates the r − k class estimator for values of k such that .
(ii) If for all i ∈ Nr and , then the r − (k, d) class estimator dominates the r − k class estimator for all values of k > 0.
(iii) If for all i ∈ Nr, the r − k class estimator dominates the r − (k, d) class estimator for all values of k > 0 and 0 < d < 1.
The MSE of the two-parameter class estimator can be obtained by substituting r = p in (2.10), given as:
From (2.10) and (3.26), the difference in the MSEs, denoted as Δ5, is given by:
which can be further simplified as:From (3.27), it is evident that Δ5 is positive if
If for all i ∈ Np−r, Δ5 is positive when for all i ∈ Np−r, i.e.
for all values of k > 0.However, when
Δ5 is positive for the values of k such thatFurthermore, if for all i ∈ Np−r and d satisfies (3.29), Δ5 is positive for the values of k, which satisfies
And, if d satisfies (3.30), Δ5 is negative for all values of k > 0 and 0 < d < 1.The comparisons can be concluded in the following theorem.
(i) If for all i ∈ Np−r and , then Δ5 > 0 for all k > 0.
(ii) If for all i ∈ Np−r and 0 < d < , then Δ5 > 0 when k is such that
(iii) If for all i ∈ Np−r and , then Δ5 > 0 when k is such that .
(iv) If for all i ∈ Np−r and , then Δ5 < 0 for all values of k > 0.
In this section, conditions for dominance of the r − (k, d) class estimator over the OLS, ORR, PCR, r − k class estimator and the two-parameter class estimator under the MSE criterion in the misspecified model have been obtained. However, the range of dominance does not remain the same in the misspecified model as it is in the model assumed to be correct. Moreover, the depletion or enlargement of the dominance range for the r − (k, d) class estimator over the other competing estimators depend on certain parametric conditions. For instance, if for all i ∈ N, the range of dominance of the r − (k, d) class estimator over the OLS estimator increases in the misspecified model. Furthermore, a Monte Carlo study has been conducted to understand the effect of misspecification on the dominance of the r − (k, d) class estimator over the other competing estimators.
To compare the dominance of the estimators in true (when there is no misspecification in the model) and misspecified model, the regressors have been generated by the method given in McDonald and Galarneau (1975) and Gibbons(1981), which is defined as:
where wi and wj are n × 1 vectors of independent standard normal pseudo-random numbers, ρ is specified so that the correlation between any two regressors is given by ρ2. The dependent variable y has been generated as follows: where ζ = [βγ]. u is a vector of normal pseudo-random numbers with standard deviation σ. Following McDonald Galarneau (1975), Gibbons (1981), Kibria (2003) and others, ζ has been chosen as the normalized eigenvector corresponding to the largest eigenvalue of the D′D matrix. As this study is aimed at studying the effect of the omission of relevant regressors on the performance of some competing estimators of β, the following is estimated for the model (4.1) and the misspecified model: when there is no misspecification, both X and Z have been used in estimation and when there is misspecification due to the omission of relevant regressors, information in Z matrix has not been used to estimate β. For example, the OLS estimator for the misspecified model is obtained by: and the OLS estimate of β in the case of no misspecification is obtained by taking first p components of the OLS estimate of ζ, given as:In this study, simulation is done for some selected values of n, p, q, ρ, σ2, k and d to compare the performance of the estimators. The values of the parameters are taken as: n = 50; p = 5; q = 3; ρ = 0.95,0.99; σ = 0.5,1; k = 0.1,0.5,0.9,1.5,5 and d = 0.1,0.5,0.9. The value of r is decided by a scree plot, which is drawn between eigenvalues and components (see Johnson and Wichern (2007)). For each parametric combination, the simulation process has been repeated 2500 times and the estimated MSE (EMSE) is calculated by the following formula
where is the estimated value of β in ith iteration. The results of the simulation are shown in Tables 1 to 4, where EMSE of estimators in true model and in misspecified model are denoted by EMSET and EMSEM, respectively, and , , , , and denote the OLS, ORR, PCR, r − k class, two-parameter class and r − (k, d) class estimators respectively. The following remarks are made from simulation results:Since ρ affects the structure of the design matrix, the estimated MSEs of and are the same for all values of k and d for a fixed σ in true and misspecified models. As expected, for higher value of σ, the estimated MSEs inflate for all the estimators in true and misspecified model as well. Similarly, when the collinearity among the regressors increases, the estimated MSEs of the estimators inflate in both the models.
As the theoretical results suggest, the MSE of the estimator may increase due to the omission of relevant regressors depending on the values of unknown parameters. When we compare the performances of the estimators in true and misspecified model for all choices of the parameters involved almost all the estimators have larger estimated MSE in the misspecified model than in the case where there is no misspecification.
While examining the variations in the estimated MSE of the estimators with respect to the variations in k and d from Tables 1-4, we observe that as the value of k increases, the values of the estimated MSEs decrease for all the estimators considered here where k is involved. However, in true model exhibits a pattern of concave up function of k, that is the estimated MSE of first decreases and then increases after attaining a minimum value of the MSE with the increase in the value of k. In our simulation, the minimum value of the MSE of the r − (k, d) class estimator when d = 0.1,0.5 is attained for some value of k in between 0.9 to 1.5 and 1.5 to 5 for σ = 0.5 and σ = 1 respectively.
However, with the increase in the value of d, the estimated MSEs of and increase for the selected values of ρ and σ for both the models. The values of the estimated MSEs show that the r − (k, d) class estimator performs better than the OLS, ORR, PCR, two-parameter class estimator for all chosen values of k, d, σ and ρ, although the dominance of the r − (k, d) class estimator over the r − k class estimator depends on the choices of k and d. In fact, the difference in the estimated MSE values of the r − (k, d) class estimator and r − k class estimator do not show much difference if seen up to the third or forth decimal places for small σ, however, if observed up to the sixth or seventh decimal places, the MSE of the r − k class estimator is found to be less than that of the r − (k, d) class estimator. For σ = 1, the r − k class estimator shows dominance over the r − (k, d) class estimator in the misspecified model, see Table 2 and 4, the reason being the condition of dominance of the r − (k, d) class estimator over the r − k class estimator (see Theorem 3.4) is not satisfied in this simulation.
In order to illustrate our theoretical results, in this section we now consider the data set on Total National Research and Development Expenditures as a Per cent of Gross National Product originally due to Gruber (1998), also analysed by Zhong and Yang (2007). It represents the relationship between the dependent variable Y, the percent spent by the U.S., and the four other independent variables X1, X2, X3 and X4. The variables X1, X2, X3 and X4, respectively represents the percent spent by France, West Germany, Japan and the former Soviet Union. The variables are standardized and the OLS estimator of β = (β1 β2 β3 β4)′ is obtained as β = (0.6455,0.0896,0.1436,0.1526)'. We obtain the eigenvalues of X′X as λ1 = 302.9626, λ2 = 0.7283, λ3 = 0.0446, and λ4 = 0.0345, and the condition number is approximately 8,776.382. Hence, the design matrix is quite ill-conditioned.
Now, let us consider that the investigator has omitted Z = [X4] mistakenly, which results in the misspecified model (2.2) with X matrix having 3 variables X1, X2, and X3. The eigenvalues of the X matrix in the misspecified model are 161.38584077, 0.10961836 and 0.04454088, and the condition number is 3623.32, which indicates an ill-conditioned design matrix in the misspecified model. The OLS estimators of β, γ and σ2 in the model (2.2) is obtained as β = (0.80878236, 0.41402294, −0.09630492)', , respectively and we chose r = 2. The values of k and d are chosen as: k0 = , where αmax is the maximum element of α = T’β, which was suggested by Hoerl and Kennard (1970) and d = 0.0557 is the positive solution of . The MSEs of the estimators are estimated by replacing β with the PCR estimator, which is an unbiased estimator, and are presented in Table 5 along with the estimated values of regression coefficients for both true and misspecified model. Figure 2 represents the estimated MSEs of the estimators in the two models.
From Table 5, we can see the sign of β3 has changed in the misspecified model from positive to negative, which gives an evidence of the well-established results that the omission affects the estimation of parameters. Further, the estimated MSEs increase in the misspecified model as compared to the true model. We observe that the r−(k, d) class estimator outperforms the OLS, ORR, two-parameter class and PCR estimators in MSE sense. However, the MSEs of the r-k class estimator and the r-(k, d) class estimator are almost equal and the difference can be only noticed at sixth decimal place. The dominance of the estimators can be easily seen in Figure 2.
On the other hand, from the results stated in Table 5 for the misspecified model, we see that the r−(k, d) class estimator is superior to the OLS, ORR, two-parameter class and PCR estimators, and does not perform better than the r −k class estimator under the MSE criterion. Moreover, the theoretical findings obtained in this study support the numerical results given in Table 5. Now, in order to verify the conditions of the dominance under MSE criterion, let us take Theorem 3.1, where we get , −0.0042, 0.0027, clearly condition (ii) of the theorem will be applied and the lower limits of d are −0.1592445−1.9733796, thus the r−(k, d) class estimator dominates the OLS estimator for all values of d, which is the result obtained in the numerical illustration. Next, let us take the condition of dominance of r−(k, d) class over the r −k class given in Theorem 3.4; , 0.0037 for i = 1,2 and the value of d is 0.0557, which satisfies the condition (i) in Theorem 3.4 as the values of for i=1,2 are 1.0856, 0.1421. Further, the value of the lower bound of k in condition (i) of Theorem 3.4 comes out to be 63.872591. Evidently the condition is not satisfied, hence the r−(k, d) class estimator does not dominate the r −k class estimator in this numerical illustration. Similarly, other conditions can also be verified.
In this paper the effect of misspecification due to omission of relevant regressors in a linear regression model when the problem of multicollinearity exists, on the dominance of the r-(k,d) class estimator over the other competing estimators have been studied. The dominance conditions of the r-(k,d) class estimator over the OLS, ORR, PCR, r-k class and the two-parameter class estimators have been derived under scalar mean squared error criterion. It has been observed that the MSE of the estimators may increase or decrease due to misspecification depending on the values of the unknown parameters. Similarly, the ranges of dominance of the r-(k,d) class estimator over the others may shrink or widen in the misspecified model. To understand the effect of misspecification on dominance of the r-(k,d) class estimator over the others a Monte Carlo simulation and a numerical example have been given and it is observed that the MSE of the estimators increases in the misspecified model as compared to the model assumed to be true. The r-(k,d) class estimator performs better than the OLS, ORR, two-parameter class estimator and the PCR estimator in the misspecified model as well for all chosen values of the parameters. However, the r-(k,d) class estimator and the r-k class estimator do equally well when observed up to few decimal places in simulation, whereas in the numerical example the r-k class estimator is found to be the most suited as an alternative to the OLS estimator in the misspecified model with multicollinearity. Hence, the study stuggests that the r-k class estimator or the r-(k,d) class estimators are a better choice over the other estimators considered in this study in the case of the misspecified model with multicollinearity.