7
Partially Linear Regression Models

7.1 Introduction

In this chapter, the problem of ridge estimation is studied in the context of partially linear models (PLMs). In a nutshell, PLMs are smoothed models that include both parametric and nonparametric parts. They allow more flexibility compared to full/nonparametric regression models.

Consider the usual PLM with the form

(7.1)equation

where images is a vector of explanatory variables, images is an unknown images‐dimensional parameter vector, the images's are known and nonrandom in some bounded domain images, images is an unknown smooth function, and images's are i.i.d. random errors with mean 0, variance images, which are independent of images. PLMs are more flexible than standard linear models since they have both parametric and nonparametric components. They can be a suitable choice when one suspects that the response images linearly depends on images, but that it is nonlinearly related to images.

Surveys regarding the estimation and application of the model (7.1) can be found in the monograph of Hardle et al. (2000). Raheem et al. (2012) considered absolute penalty and shrinkage estimators in PLMs where the vector of coefficients images in the linear part can be partitioned as images; images is the coefficient vector of the main effects, and images is the vector of the nuisance effects. For a more recent study about PLM, we refer to Roozbeh and Arashi (2016b). Since for estimation images, we need to estimate the nonparametric component images, we estimate it using the kernel smoothing method. Throughout, we do not further discuss the estimation of images in PLMs, since the main concern is the estimation of images in the ridge context.

To estimate images, assume that images, images satisfy the model (7.1). Since images, we have images for images. Hence, if we know images, a natural nonparametric estimator of images is given by

(7.2)equation

where the positive weight function images satisfies the three regularity conditions given here:

  1. images,
  2. images,
  3. images, where images is the indicator function of the set images,
    equation

These assumptions guarantee the existence of images at the optimal convergence rate images, in PLMs with probability one. See Müller and Rönz (1999) for more details.

7.2 Partial Linear Model and Estimation

Consider the PLM defined by

(7.3)equation

where images's are responses, images, images, are covariates of the unknown vectors, namely, images and also images, respectively, and images are design points, images is a real valued function defined on images, and images are independently and identically distributed (i.i.d.) errors with zero mean and constant variance, images.

Our main objective is to estimate images when it is suspected that the images‐dimensional sparsity condition images may hold. In this situation, we first estimate the parameter images using's Speckman's (1988) approach of partial kernel smoothing method, which attains the usual parametric convergence rate images without undersmoothing the nonparametric component images.

Assume that images satisfy the model (7.3). If images is the true parameter, then images. Also, for images, we have

equation

Thus, a natural nonparametric estimator of images given images is

(7.4)equation

with the probability weight functions images satisfies the three regularity conditions (i)–(iii).

Suppose that we partitioned the design matrix,

equation

and others as

equation

where

equation

To estimate images, we minimize

(7.5)equation

with respect to images. This yields the estimator of images as

(7.6)equation
(7.7)equation

where

equation

respectively.

If images is true, then the restricted estimator is

(7.8)equation

Further, we assume that images are i.i.d. random variables and let

equation

Then, the conditional variance of images given images is images, where images with images. Then, by law of large numbers (LLN),

equation

and the matrix images is nonsingular with probability 1. Readers are referred to Shi and Lau (2000), Gao (1995, 1997), and Liang and Härdle (1999), among others. Moreover, for any permutation images of images as images,

(7.9)equation

Finally, we use the following assumption:

The functions images and images satisfy the Lipchitz condition of order 1 on images for images. Then, we have the following theorem.

As a result of Theorem 7.1, we have the following limiting cases:

  1. images
  2. images
  3. images where images is the imagesth diagonal element of images.

7.3 Ridge Estimators of Regression Parameter

Let the PLM be given as the following form

(7.10)equation

where

equation

In general, we assume that images is an images‐vector of i.i.d. random variables with distribution images, where images is a symmetric, positive definite known matrix and images is an unknown parameter.

Using the model (7.20) and the error distribution of images we obtain the maximum likelihood estimator (MLE) or generalized least squares estimator (GLSE) of images by minimizing

(7.11)equation

to get

(7.12)equation

where images, images, images, and images, images.

Now, under a restricted modeling paradigm, suppose that images satisfies the following linear nonstochastic constraints

(7.13)equation

where images is a images nonzero matrix with rank images and images is a images vector of prespecified values. We refer restricted partially linear model (RPLM) to (7.20). For the RPLM, one generally adopts the well‐known GLSE given by

(7.14)equation

The GLSE is widely used as an unbiased estimator. We refer to Saleh (2006) and Saleh et al. (2014) for more details and applications of restricted parametric or nonparametric models.

Now, in the line of this book, if there exists multicollinearity in images, the images would be badly apart from the actual coefficient parameter in some directions of images‐dimension space. Hence, instead of minimizing the GLSE objective function as in (7.21), following Roozbeh (2015), one may minimize the objective function where both images and images are penalty functions

(7.15)equation

where images is a vector of constants.

The resulting estimator is the restricted generalized ridge estimator (GRE), given by

(7.16)equation

where

(7.17)equation

is the GRE and images is the ridge parameter.

From Saleh (2006), the likelihood ratio criterion for testing the null hypothesis images, is given by

(7.18)equation

where images, images, is an unbiased estimator of images.

Note that images follows a noncentral images‐distribution with images DF and noncentrality parameter images given by

(7.19)equation

Following Saleh (2006), we define three sorts of estimators using the test statistic images. First, we consider the preliminary test generalized ridge estimator (PTGRE) defined by

(7.20)equation

where images is the upper images‐level critical value for the test of images and images is the indicator function of the set images.

This estimator has been considered by Saleh and Kibria (1993). The PTGRE has the disadvantage that it depends on images, the level of significance, and also it yields the extreme results, namely, images and images depending on the outcome of the test.

A nonconvex continuous version of the PTGRE is the Stein‐type generalized ridge estimator (SGRE) defined by

(7.21)equation

The Stein‐type generalized ridge regression has the disadvantage that it has strange behavior for small values of images. Also, the shrinkage factor images becomes negative for images. Hence, we define the positive‐rule Stein‐type generalized ridge estimator (PRSGRE) given by

(7.22)equation

Shrinkage estimators have been considered by Arashi and Tabatabaey (2009), Arashi et al. (2010, 2012), Arashi (2012), and extended to monotone functional estimation in multidimensional models just as the additive regression model, semi‐parametric PLM, and generalized linear model by Zhang et al. (2008).

7.4 Biases and images Risks of Shrinkage Estimators

In this section, we give exact expressions of the bias and images‐risk functions for the estimators images and images. Since the properties of unrestricted, restricted, and preliminary test estimators have been widely investigated in the literature, we refer to Sarkar (1992) and Saleh and Kibria (1993) for the properties of unrestricted, restricted, and preliminary test generalized ridge regression estimator.

For analytical comparison between the proposed estimators, we refer to Roozbeh (2015) and continue our study with numerical results.

7.5 Numerical Analysis

To examine the images‐risk performance of the proposed estimators, we adopted the Monte Carlo simulation study of Roozbeh (2015) for illustration.

To achieve different degrees of collinearity, following Kibria (2003) the explanatory variables were generated using the given model for images:

equation

where images are independent standard normal pseudorandom numbers, and images is specified so that the correlation between any two explanatory variables is given by images. These variables are then standardized so that images and images are in correlation forms. Three different sets of correlation corresponding to images, and 0.99 are considered. Then images observations for the dependent variable are determined by

(7.24)equation

where images, and

equation

which is a mixture of normal densities for images and images is the normal probability density function (p.d.f.) with mean images and variance images. The main reason for selecting such a structure for the nonlinear part is to check the efficiency of the nonparametric estimations for wavy function. This function is difficult to estimate and provides a good test case for the nonparametric regression method.

According to Roozbeh (2015), it is assumed images, for which the elements of images are images and images. The set of linear restrictions is assumed to have form images, where

equation

For the weight function images, we use

equation

which is Priestley and Chao's weight with the Gaussian kernel. We also apply the cross‐validation (CV) method to select the optimal bandwidth images, which minimizes the following CV function

equation

where images obtains by replacing images and images by

equation

Here, images is the predicted value of images at images with images and images left out of the estimation of the images.

The ratio of largest eigenvalue to smallest eigenvalue of the design matrix in model (7.24) is approximately images, and 14 495.28 for images, and 0.99, respectively.

In Tables 7.17.6, we computed the proposed estimators' images risk,

equation

for different values of images and images.

We found the best values of images (images) by plotting

equation

vs. images for each of the proposed estimators (see Figure 7.1).

Figure 7.2 shows the fitted function by kernel smoothing after estimation of the linear part of the model using images and images, that are, images, respectively, for images, and 0.99. The minimum of CV approximately obtained was images for the model (7.24) with images. The diagram of CV vs. images is also plotted in Figure 7.3.

From the simulation, we found that at most images for superiority of PRSGRE over SGLSE are 0.598, 0.792, and 0.844 for images, and 0.99, respectively. It is realized that the ridge parameter images is an increasing function of images for superiority of PRSGRE over SGRE in images‐risk sense. These images values are 0.527, 0.697, and 0.730 for superiority of PRSGRE over PRSGLSE. Finally, the PRSGRE is better than SGRE for all values of images and images in images‐risk sense.

Table 7.1 Evaluation of the Stein‐type generalized RRE at different images values in model (7.24) with images.

Coefficients (images) 0 0.05 0.10 0.15 0.20 0.25 images = 0.299 0.30
images images1.007 images1.009 images1.010 images1.012 images1.014 images1.015 images0.994 images1.017
images images0.900 images0.902 images0.903 images0.905 images0.906 images0.908 images0.995 images0.909
images   1.986   1.986   1.986   1.986   1.987   1.987   2.004   1.987
images   3.024   3.025   3.026   3.027   3.028   3.029   3.004   3.030
images images5.073 images5.075 images5.078 images5.081 images5.083 images5.085 images5.002 images5.087
images   3.999   4.002   4.005   4.008   4.011   4.013   3.996   4.016
images   0.274   0.274   0.273   0.273   0.273   0.273   0.273   0.273

Table 7.2 Evaluation of PRSGRE at different images values in model (7.24) with images.

Coefficients (images) 0 0.05 0.10 0.15 0.20 0.25 images = 0.2635 0.30
images images1.004 images1.002 images1.000 images0.998 images0.996 images0.994 images0.993 images0.992
images images1.004 images1.002 images1.000 images0.999 images0.997 images0.995 images0.995 images0.994
images   2.008   2.007   2.006   2.006   2.005   2.005   2.004   2.004
images   3.012   3.010   3.009   3.007   3.006   3.004   3.004   3.003
images images5.020 images5.016 images5.013 images5.009 images5.006 images5.003 images5.0021 images4.999
images   4.016   4.012   4.008   4.004   4.000   3.997   3.996   3.993
images   0.242   0.241   0.241   0.241   0.241   0.241   0.241   0.241
images   0.027   0.027   0.027   0.028   0.028   0.028   0.028   0.028

Table 7.3 Evaluation of SGRE at different images values in model (7.24) with images.

Coefficients (images) 0 0.10 0.20 0.30 images = 0.396 0.40 0.50 0.60
images images0.969 images0.968 images0.966 images0.965 images0.964 images0.964 images0.962 images0.961
images images0.968 images0.967 images0.966 images0.964 images0.963 images0.963 images0.962 images0.961
images   2.025   2.024   2.024   2.0232   2.022   2.022   2.021   2.020
images   3.009   3.008   3.006   3.005   3.003   3.003   3.002   3.001
images images5.061 images5.058 images5.055 images5.052 184 images5.049 images5.048 images5.045 images5.042
images   3.992   3.989   3.985   3.982 498   3.979   3.979   3.975   3.972
images   0.540   0.537   0.535   0.534   0.534   0.534   0.534   0.536
images   0.090   0.090   0.089   0.089   0.089   0.089   0.088   0.088

Table 7.4 Evaluation of PRSGRE at different images values in model (7.24) with images.

Coefficients (images) 0 0.10 0.20 0.30 images = 0.348 0.40 0.50 0.60
images images0.966 images0.963 images0.956 images0.950 images0.947 images0.943 images0.937 images0.931
images images0.968 images0.963 images0.957 images0.952 images0.949 images0.946 images0.941 images0.936
images   2.025   2.023   2.021   2.018   2.017   2.016   2.014   2.012
images   3.009   3.005   3.000   2.996   2.994   2.991   2.987   2.983
images images5.061 images5.050 images5.039 images5.027 images5.022 images5.016 images5.005 images4.994
images   3.992   3.978   3.965   3.951   3.944   3.937   3.924   3.911
images   0.476   0.474   0.472   0.471 787   0.471   0.471   0.472   0.474
images   0.0903   0.0905   0.090   0.090   0.090   0.091   0.0912   0.091

Table 7.5 Evaluation of SGRE at different images values in model (7.24) with images.

Coefficients (images) 0 0.10 0.20 0.30 0.40 images = 0.422 0.50 0.60
images images1.430 images1.366 images1.308 images1.257 images1.210 images1.200 images1.167 images1.127
images images1.267 images1.205 images1.150 images1.100 images1.056 images1.046 images1.015 images0.977
images   1.487   1.500   1.510   1.517   1.522   1.523   1.525   1.527
images   3.046   3.024   3.001   2.977   2.954   2.949   2.930   2.906
images images4.945 images4.876 images4.811 images4.747 images4.687 images4.674 images4.628 images4.572
images   5.025   4.854   4.700   4.563   4.438   4.412   4.324   4.221
images   5.748   5.391   5.179   5.082   5.076   5.085   5.142   5.265
images   0.040   0.040   0.041   0.041   0.042   0.042   0.042   0.042

Table 7.6 Evaluation of PRSGRE at different images values in model (7.24) with images.

Coefficients (images) 0 0.10 0.20 0.30 images = 0.365 0.40 0.50 0.60
images images1.430 images1.340 images1.260 images1.188 images1.144 images1.122 images1.062 images1.007
images images1.267 images1.179 images1.101 images1.030 images0.987 images0.965 images0.907 images0.854
images   1.487   1.496   1.501   1.503   1.503   1.503   1.500   1.496
images   3.046   3.007   2.967   2.927   2.901   2.887   2.847   2.808
images images4.945 images4.834 images4.728 images4.626 images4.562 images4.528 images4.433 images4.342
images   5.025   4.786   4.572   4.379   4.264   4.205   4.046   3.901
images   5.060   4.752   4.584   4.525   4.534   4.552   4.648   4.798
images   0.040   0.041   0.042   0.042   0.043   0.043   0.044   0.045

7.5.1 Example: Housing Prices Data

The housing prices data consist of 92 detached homes in the Ottawa area that were sold during 1987. The variables are defined as follows. The dependent variable is sale price (SP), the independent variables include lot size (lot area; LT), square footage of housing (SFH), average neighborhood income (ANI), distance to highway (DHW), presence of garage (GAR), and fireplace (FP). The full parametric model has the form

(7.25)equation

Looking at the correlation matrix given in Table 7.7, there exists a potential multicollinearity between variables SFH & FP and DHW & ANI.

We find that the eigenvalues of the matrix images are given by images, images, images, images, images, images, and images. It is easy to see that the condition number is approximately equal to images. So, images is morbidity badly.

Plots depicting D and risks versus k for different values of a of the proposed estimators.

Figure 7.1 Plots of images and risks vs. images for different values of images.

Graphs depicting the estimation of the fitted functions of normal p.d.fs by the kernel approach. Solid lines are the estimates and dashed lines are the true functions.

Figure 7.2 Estimation of the mixtures of normal p.d.fs by the kernel approach. Solid lines are the estimates and dotted lines are the true functions.

Graphical curve depicting the plot of cross-validation function versus hn.

Figure 7.3 Plot of CV vs. images.

Table 7.7 Correlation matrix.

Variable SP LT SFH FP DHW GAR ANI
SP   1.00 0.14 0.47 0.33 images0.10 0.29   0.34
LT   0.14 1.00 0.15 0.15   0.08 0.16   0.13
SFH   0.47 0.15 1.00 0.46   0.02 0.22   0.27
FP   0.33 0.15 0.46 1.00   0.10 0.14   0.38
DHW images0.10 0.08 0.02 0.10   1.00 0.05 images0.10
GAR   0.29 0.16 0.22 0.14   0.05 1.00   0.02
ANI   0.34 0.13 0.27 0.38 images0.10 0.02   1.00

An appropriate approach, as a remedy for multicollinearity, is to replace the pure parametric model with a PLM. Akdeniz and Tabakan (2009) proposed a PLM (here termed as semi‐parametric model) for this data by taking the lot area as the nonparametric component.

Plots depicting individual explanatory variables versus dependent variable, linear fit (dashed line) and local polynomial fit (solid line).

Figure 7.4 Plots of individual explanatory variables vs. dependent variable, linear fit (dash line), and local polynomial fit (solid line).

To realize the type of relation from linearity/nonlinearity viewpoint between dependent variable (SP) and explanatory variables (except for the binary ones), they are plotted in Figure 7.4. According to this figure, we consider the average neighborhood income (ANI) as a nonparametric part. So, the specification of the semi‐parametric model is

(7.26)equation

To compare the performance of the proposed restricted estimators, following Roozbeh (2015), we consider the parametric restriction images, where

equation

The test statistic for testing images, given our observations, is

equation

where images. Thus, we conclude that the null‐hypothesis images is accepted.

Table 7.8 summarizes the results. The “parametric estimates” refer to a model in which ANI enters. In the “semiparametric estimates,” we have used the kernel regression procedure with optimal bandwidth images for estimating images. For estimating the nonparametric effect, first we estimated the parametric effects and then applied the kernel approach to fit images on images for images, where images (Figure 7.5).

Table 7.8 Fitting of parametric and semi‐parametric models to housing prices data.

Parametric estimates Semiparametric estimates
Variable Coef. s.e. Coef. s.e.
Intercept 62.67 20.69
LT 0.81  2.65 9.40  2.25
SFH 39.27 11.99 73.45 11.15
FP 6.05  7.68 5.33  8.14
DHW    images8.21  6.77    images2.22  7.28
GAR 14.47  6.38 12.98  7.20
ANI 0.56  0.28
images 798.0625 149.73
images 33 860.06 88 236.41
images 0.3329 0.86

The ratio of largest to smallest eigenvalues for the new design matrix in model (7.25) is approximately images, and so there exists a potential multicollinearity between the columns of the design matrix. Now, in order to overcome the multicollinearity for better performance of the estimators, we used the proposed estimators for model (7.25).

The SGRRE and PRSGRE for different values of the ridge parameter are given in Tables 7.9 and 7.10, respectively. As it can be seen, images is the best estimator for the linear part of the semi‐parametric regression model in the sense of risk.

Graph depicting the estimation of nonlinear effect of ANI on dependent variable by kernel fit.

Figure 7.5 Estimation of nonlinear effect of ANI on dependent variable by kernel fit.

In Figure 7.6, the images and images of SGRE (solid line) and PRSGDRE (dash line) vs. ridge parameter images are plotted. Finally, we estimated the nonparametric effect (images) after estimation of the linear part by images in Figure 7.7, i.e. we used kernel fit to regress images on images.

Table 7.9 Evaluation of SGRRE at different images values for housing prices data.

Variable (images) 0.0 0.10 0.20 images = 0.216 0.30 0.40 0.50 0.60
Intercept
LT    9.043    9.139    9.235    9.331    9.427    9.250    9.524    9.620
SFH   84.002   84.106   84.195   84.270   84.334   84.208   84.385   84.425
FP    0.460    0.036    0.376    0.777    1.168    0.441    1.549    1.919
DHW images4.086 images4.382 images4.673 images4.960 images5.241 images4.719 images5.519 images5.792
GAR    4.924    4.734    4.548    4.367    4.190    4.519    4.017    3.848
ANI
images  257.89  254.09  252.62  252.59  253.35  256.12  260.82  267.31
images 2241.81 2241.08 2240.50 2240.42 2240.12 2240.03 2240.02 2240.18

Table 7.10 Evaluation of PRSGRRE at different images values for housing prices data.

Variable (images) 0.0 0.10 images = 0.195 0.20 0.30 0.40 0.50 0.60
Intercept
LT    9.226    9.355    9.476    9.481    9.604    9.725    9.843    9.958
SFH   78.614   77.705   76.860   76.820   75.959   75.120   74.303   73.506
FP    2.953    3.229    3.483    3.495    3.751    3.998    4.237    4.466
DHW images3.136 images3.033 images2.937 images2.933 images2.835 images2.739 images2.645 images2.554
GAR    9.042    9.112    9.176    9.179    9.242    9.303    9.362    9.417
ANI
images  234.02  230.68  229.67  230.82  232.68  234.02  239.12  246.00
images 2225.762 2233.685 2240.480 2240.838 2248.805 2256.785 2264.771 2272.758
Graphical curves depicting the D and risk of SGRE (solid line) and PRSGDRE (dash line) versus ridge parameter k.

Figure 7.6 The diagrams of images and risk vs. images for housing prices data.

Graphs depicting the estimation of f(ANI) by kernel regression after removing the linear part by proposed estimators in housing prices data.

Figure 7.7 Estimation of images(ANI) by kernel regression after removing the linear part by the proposed estimators in housing prices data.

7.6 High‐Dimensional PLM

In this section, we develop the theory proposed in the previous section for the high‐dimensional case in which images. For our purpose, we adopt the structure of Arashi and Roozbeh (2016), i.e. we partition the regression parameter images as images, where the sub‐vector images has dimension images, images and images with images. Hence, the model (7.20) has the form

(7.27)equation

where images is partitioned according to images in such a way that images is a images sub‐matrix, images.

In this section, we consider the estimation of images under the sparsity assumption images. Hence, we minimize the objective function

(7.28)equation

where images is a function of sample size images and images is a vector of constants.

The resulting estimator of images is the generalized ridge regression estimator (RRE), given by

(7.29)equation

where images and images is the generalized restricted estimator of images given by

(7.30)equation

Here, images is the bandwidth parameter as a function of sample size images, where we estimate images by images, images with images is a kernel function of order images with bandwidth parameter images (see Arashi and Roozbeh 2016).

Note that the generalized (unrestricted) ridge estimators of images and images, respectively, have the forms

(7.31)equation

where

equation

In order to derive the forms of shrinkage estimators, we need to find the test statistic for testing the null‐hypothesis images. Thus, we make use of the following statistic

(7.32)equation

where images and

(7.33)equation

To find the asymptotic distribution of images, we need to make some regularity conditions. Hence, we suppose the following assumptions hold

  1. (A1) (A1) images, where images is the imagesth row of images.
  2. (A2) (A2) images as images,
  3. (A3) (A3) Let
    equation

    Then, there exists a positive definite matrix images such that

    equation
  4. (A4) (A4) images.
  5. (A5) (A5) images, images.

Note that by (A2), (A4), and (A5), one can directly conclude that as images

(7.34)equation

The test statistic images diverges as images, under any fixed alternatives images. To overcome this difficulty, in sequel, we consider the local alternatives

equation

where images is a fixed vector.

Then, we have the following result about the asymptotic distribution of images. The proof is left as an exercise.

Then, in a similar manner to that in Section 5.2, the PTGRE, SGRE, and PRSGRE of images are defined respectively as

(7.35)equation
(7.36)equation

Under the foregoing regularity conditions (A1)–(A5) and local alternatives images, one can derive asymptotic properties of the proposed estimators, which are left as exercise. To end this section, we present in the next section a real data example to illustrate the usefulness of the suggested estimators for high‐dimensional data.

7.6.1 Example: Riboflavin Data

Here, we consider the data set about riboflavin (vitamin images) production in images, which can be found in the R package “hdi”. There is a single real‐valued response variable which is the logarithm of the riboflavin production rate. Furthermore, there are images explanatory variables measuring the logarithm of the expression level of 4088 genes. There is one rather homogeneous data set from images samples that were hybridized repeatedly during a fed batch fermentation process where different engineered strains and strains grown under different fermentation conditions were analyzed. Based on 100‐fold cross‐validation, the LASSO shrinks 4047 parameters to zero and remains images significant explanatory variables.

To detect the nonparametric part of the model, for images, we calculate

equation

where images is obtained by deleting the imagesth column of matrix images. Among all 41 remaining genes, “images” had minimum images value. We also use the added‐variable plots to identify the parametric and nonparametric components of the model. Added‐variable plots enable us to visually assess the effect of each predictor, having adjusted for the effects of the other predictors. By looking at the added‐variable plot (Figure 7.8), we consider “images” as a nonparametric part. As it can be seen from this figure, the slope of the least squares line is equal to zero, approximately, and so, the specification of the sparse semi‐parametric regression model is

(7.37)equation

where images and images.

Graph depicting added variable plot of explanatory variables DNAJ_at versus dependent variable, linear fit (solid line), and kernel fit (dashed line).

Figure 7.8 Added‐variable plot of explanatory variables images vs. dependent variable, linear fit (solid line) and kernel fit (dashed line).

For estimating the nonparametric part of the model, images, we use

equation

which is Priestley and Chao's weight with the Gaussian kernel.

We also apply the CV method to select the optimal bandwidth images and images, which minimizes the following CV function

equation

where images is obtained by replacing images and images with images, images, images, images, images, images.

Table 7.11 shows a summary of the results. In this table, the RSS and images, respectively, are the residual sum of squares and coefficient of determination of the model, i.e. images, and images. For estimation of the nonparametric effect, at first we estimated the parametric effects by one of the proposed methods and then the local polynomial approach was applied to fit images on images, images.

Table 7.11 Evaluation of proposed estimators for real data set.

Method GRE GRRE SGRRE PRSGRRE
RSS 16.1187 1.9231 1.3109 1.1509
images  0.7282 0.9676 0.9779 0.9806

In this example, as can be seen from Figure 7.8, the nonlinear relation between dependent variable and images can be detected, and so the pure parametric model does not fit the data; the semi‐parametric regression model fits more significantly. Further, from Table 7.11, it can be deduced that PRSGRE is quite efficient, in the sense that it has significant value of goodness of fit.

7.7 Summary and Concluding Remarks

In this chapter, apart from the preliminary test and penalty estimators, we mainly concentrated on the Stein‐type shrinkage ridge estimators, for estimating the regression parameters of the partially linear regression models. We developed the problem of ridge estimation for the low/high‐dimensional case. We analyzed the performance of the estimators extensively in Monte Carlo simulations, where the nonparametric component of the model was estimated by the kernel method. The housing prices data used the low‐dimensional part and the riboflavin data used for the high‐dimensional part for practical illustrations, in which the cross‐validation function was applied to obtain the optimal bandwidth parameters.

Problems

  1. 7.1 Prove that the test statistic for testing images is
    equation

    and images is the images‐level critical value. What would be the distribution of images under both null and alternative hypotheses?

  2. 7.2 Prove that the biases of the Stein‐type and positive‐rule Stein‐type of generalized RRE are, respectively, given by
    equation

    where images and images are defined in Theorem 7.2.

  3. 7.3 Show that the risks of the Stein‐type and positive‐rule Stein‐type of generalized RRE are, respectively, given by
    equation
  4. 7.4 Prove that the test statistic for testing images is
    equation

    where images and

    equation
  5. 7.5 Prove Theorem 7.2.
  6. 7.6 Consider a real data set, where the design matrix elements are moderate to highly correlated, then find the efficiency of the estimators using unweighted risk functions. Find parallel formulas for the efficiency expressions and compare the results with that of the efficiency using weighted risk function. Are the two results consistent?
..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset