The NLIN Procedure

Measures of Nonlinearity, Diagnostics and Inference

A "close-to-linear" nonlinear regression model, in the sense of Ratkowsky (1983, 1990), is a model in which parameter estimators have properties similar to those in a linear regression model. That is, the least squares estimators of the parameters are close to being unbiased and normally distributed, and they have minimum variance.

A nonlinear regression model sometimes fails to be close to linear due to the properties of one or several parameters. When this occurs, bias in the parameter estimates can render inferences that use the reported standard errors and confidence limits invalid.

PROC NLIN provides various measures of nonlinearity. To assess the nonlinearity of a model-data combination, you can use both of the following complementary sets of measures:

  • Box’s bias (Box 1971) and Hougaard’s skewness (Hougaard 1982, 1985) of the least squares parameter estimates

  • curvature measures of nonlinearity (Bates and Watts 1980).

Furthermore, PROC NLIN provides residual, leverage, and local-influence diagnostics (St. Laurent and Cook 1993).

In the following several sections, these nonlinearity measures and diagnostics are discussed. For this material, several basic definitions are required. Let bold upper X be the Jacobian matrix for the model, bold upper X equals StartFraction partial-differential bold f Over partial-differential bold-italic beta EndFraction, and let bold upper Q and bold upper R be the components of the QR decomposition of bold upper X equals bold upper Q bold upper R of bold upper X, where bold upper Q is an left-parenthesis n times n right-parenthesis orthogonal matrix. Finally, let bold upper B be the inverse of the matrix constructed from the first p rows of the left-parenthesis n times p right-parenthesis dimensional matrix bold upper R (that is, bold upper B equals bold upper R Subscript p Superscript negative 1). Next define

StartLayout 1st Row 1st Column left-bracket bold upper H Subscript j Baseline right-bracket Subscript k l 2nd Column equals StartFraction partial-differential squared bold f Subscript j Baseline Over partial-differential bold-italic beta Subscript k Baseline partial-differential bold-italic beta Subscript l Baseline EndFraction 2nd Row 1st Column left-bracket bold upper U Subscript j Baseline right-bracket Subscript k l 2nd Column equals sigma-summation Underscript m n Endscripts bold upper B prime Subscript k m Baseline left-bracket bold upper H Subscript j Baseline right-bracket Subscript m n Baseline bold upper B Subscript n l Baseline 3rd Row 1st Column left-bracket bold upper A Subscript j Baseline right-bracket Subscript k l 2nd Column equals StartRoot p times normal m normal s normal e EndRoot sigma-summation Underscript m Endscripts bold upper Q prime Subscript j m Baseline left-bracket bold upper U Subscript m Baseline right-bracket Subscript k l Baseline comma EndLayout

where bold upper H, bold upper U and the acceleration array bold upper A are three-dimensional left-parenthesis n times p times p right-parenthesis matrices. The first p faces of the acceleration array constitute a left-parenthesis p times p times p right-parenthesis parameter-effects array and the last left-parenthesis n minus p right-parenthesis faces constitute the left-parenthesis n minus p times p times p right-parenthesis intrinsic curvature array (Bates and Watts 1980). The previous and subsequent quantities are computed at the least squares parameter estimators.

Box’s Measure of Bias

The degree to which parameter estimators exhibit close-to-linear behavior can be assessed with Box’s bias (Box 1971) and Hougaard’s measure of skewness (Hougaard 1982, 1985). The bias and percentage bias measures are available through the BIAS option in the PROC NLIN statement. Box’s bias measure is defined as

StartLayout 1st Row 1st Column ModifyingAbove normal upper E With caret left-bracket ModifyingAbove beta With caret minus beta right-bracket 2nd Column equals minus StartFraction sigma squared Over 2 EndFraction left-parenthesis bold upper X prime bold upper W bold upper X right-parenthesis Superscript negative 1 Baseline sigma-summation Underscript i equals 1 Overscript n Endscripts w Subscript i Baseline bold x Subscript i Superscript prime Baseline trace left-parenthesis left-parenthesis bold upper X prime bold upper W bold upper X right-parenthesis Superscript negative 1 Baseline left-bracket bold upper H Subscript i Baseline right-bracket right-parenthesis EndLayout

where sigma squared equals mse if the SIGSQ option is not set. Otherwise, sigma squared is the value you set with the SIGSQ option. bold upper W is the diagonal weight matrix specified with the _WEIGHT_ variable (or the identity matrix if _WEIGHT_ is not defined) and left-bracket bold upper H Subscript i Baseline right-bracket is the left-parenthesis p times p right-parenthesis Hessian matrix at the ith observation. In the case of unweighted least squares, the bias formula can be expressed in terms of the acceleration array bold upper A,

StartLayout 1st Row 1st Column ModifyingAbove normal upper E With caret left-bracket ModifyingAbove beta With caret Subscript i Baseline minus beta Subscript i Baseline right-bracket 2nd Column equals minus StartFraction sigma squared Over 2 p times mse EndFraction sigma-summation Underscript j comma k equals 1 Overscript p Endscripts bold upper B Subscript i j Baseline left-bracket bold upper A Subscript j Baseline right-bracket Subscript k k EndLayout

As the preceding formulas illustrate, the bias depends solely on the parameter-effects array, thereby permitting its reduction through reparameterization. Example 88.4 shows how changing the parameterization of a four-parameter logistic model can reduce the bias. Ratkowsky (1983, p. 21) recommends that you consider reparameterization if the percentage bias exceeds 1%.

Hougaard’s Measure of Skewness

In addition to Box’s bias, Hougaard’s measure of skewness, g Subscript 1 i (Hougaard 1982, 1985), is also provided in PROC NLIN to assess the close-to-linear behavior of parameter estimators. This measure is available through the HOUGAARD option in the PROC NLIN statement. Hougaard’s skewness measure for the ith parameter is based on the third central moment, defined as

normal upper E left-bracket ModifyingAbove beta With caret Subscript i Baseline minus normal upper E left-parenthesis ModifyingAbove beta With caret Subscript i Baseline right-parenthesis right-bracket cubed equals minus left-parenthesis sigma squared right-parenthesis squared sigma-summation Underscript j k l Endscripts left-bracket bold upper L right-bracket Subscript i j Baseline left-bracket bold upper L right-bracket Subscript i k Baseline left-bracket bold upper L right-bracket Subscript i l Baseline left-parenthesis left-bracket bold upper V Subscript j Baseline right-bracket Subscript k l Baseline plus left-bracket bold upper V Subscript k Baseline right-bracket Subscript j l Baseline plus left-bracket bold upper V Subscript l Baseline right-bracket Subscript j k Baseline right-parenthesis

where the sum is a triple sum over the number of parameters and

bold upper L equals left-parenthesis bold upper X prime bold upper X right-parenthesis Superscript negative 1 Baseline equals left-parenthesis StartFraction partial-differential bold f Over partial-differential bold-italic beta prime EndFraction StartFraction partial-differential bold f Over partial-differential bold-italic beta EndFraction right-parenthesis Superscript negative 1

The term left-bracket bold upper L right-bracket Subscript i j denotes the value in row i, column j of the matrix bold upper L. (Hougaard (1985) uses superscript notation to denote elements in this inverse.) The matrix bold upper V is a three-dimensional left-parenthesis p times p times p right-parenthesis array

StartLayout 1st Row 1st Column left-bracket bold upper V Subscript j Baseline right-bracket Subscript k l 2nd Column equals sigma-summation Underscript m equals 1 Overscript n Endscripts StartFraction partial-differential upper F Subscript m Baseline Over partial-differential beta Subscript j Baseline EndFraction StartFraction partial-differential squared upper F Subscript m Baseline Over partial-differential beta Subscript k Baseline partial-differential beta Subscript l Baseline EndFraction EndLayout

The third central moment is then normalized using the standard error as

upper G Subscript 1 i Baseline equals normal upper E left-bracket ModifyingAbove beta With caret Subscript i Baseline minus normal upper E left-parenthesis ModifyingAbove beta With caret Subscript i Baseline right-parenthesis right-bracket cubed slash left-parenthesis sigma squared times left-bracket bold upper L right-bracket Subscript i i Baseline right-parenthesis Superscript 3 slash 2

The previous expressions depend on the unknown values of the parameters and on the residual variance sigma squared. In order to evaluate the Hougaard measure in a particular data set, the NLIN procedure computes

StartLayout 1st Row 1st Column g Subscript 1 i 2nd Column equals ModifyingAbove normal upper E With caret left-bracket ModifyingAbove beta With caret Subscript i Baseline minus normal upper E left-parenthesis ModifyingAbove beta With caret Subscript i Baseline right-parenthesis right-bracket cubed slash left-parenthesis normal m normal s normal e times left-bracket ModifyingAbove bold upper L With caret right-bracket Subscript i i Baseline right-parenthesis Superscript 3 slash 2 Baseline 2nd Row 1st Column ModifyingAbove normal upper E With caret left-bracket ModifyingAbove beta With caret Subscript i Baseline minus normal upper E left-parenthesis ModifyingAbove beta With caret Subscript i Baseline right-parenthesis right-bracket cubed 2nd Column equals minus normal m normal s normal e squared sigma-summation Underscript j k l Endscripts left-bracket ModifyingAbove bold upper L With caret right-bracket Subscript i j Baseline left-bracket ModifyingAbove bold upper L With caret right-bracket Subscript i k Baseline left-bracket ModifyingAbove bold upper L With caret right-bracket Subscript i l Baseline left-parenthesis left-bracket ModifyingAbove bold upper V With caret Subscript j Baseline right-bracket Subscript k l Baseline plus left-bracket ModifyingAbove bold upper V With caret Subscript k Baseline right-bracket Subscript j l Baseline plus left-bracket ModifyingAbove bold upper V With caret Subscript l Baseline right-bracket Subscript j k Baseline right-parenthesis EndLayout

Following Ratkowsky (1990, p. 28), the parameter beta Subscript i is considered to be very close to linear, reasonably close, skewed, or quite nonlinear according to the absolute value of the Hougaard measure StartAbsoluteValue g Subscript 1 i Baseline EndAbsoluteValue being less than 0.1, between 0.1 and 0.25, between 0.25 and 1, or greater than 1, respectively.

Relative Curvature Measures of Nonlinearity

Bates and Watts (1980) formulated the maximum parameter-effects and maximum intrinsic curvature measures of nonlinearity to assess the close-to-linear behavior of nonlinear models. Ratkowsky (1990) notes that of the two curvature components in a nonlinear model, the parameter-effects curvature is typically larger. It is this component that you can affect by changing the parameterization of a model. PROC NLIN provides these two measures of curvature both through the STATS plot-option and through the NLINMEASURES option in the PROC NLIN statement.

The maximum parameter-effects and intrinsic curvatures are defined, in a compact form, as

StartLayout 1st Row 1st Column bold upper T Superscript tau 2nd Column equals max Subscript beta Baseline StartAbsoluteValue EndAbsoluteValue beta Superscript prime Baseline bold upper A Superscript tau Baseline beta StartAbsoluteValue EndAbsoluteValue 2nd Row 1st Column bold upper T Superscript eta 2nd Column equals max Subscript beta Baseline StartAbsoluteValue EndAbsoluteValue beta Superscript prime Baseline bold upper A Superscript eta Baseline beta StartAbsoluteValue EndAbsoluteValue EndLayout

where bold upper T Superscript tau and bold upper T Superscript eta denote the maximum parameter-effects and intrinsic curvatures, while bold upper A Superscript tau and bold upper A Superscript eta stand for the parameter-effects and intrinsic curvature arrays. The maximization is carried out over a unit-vector of the parameter values (Bates and Watts 1980). In line with Bates and Watts (1980), PROC NLIN takes 10 Superscript negative 4 as the convergence tolerance for the maximum intrinsic and parameter-effects curvatures. Note that the preceding matrix products involve contraction of the faces of the three-dimensional acceleration arrays with the normalized parameter vector, beta. The corresponding expressions for the RMS (root mean square) parameter-effects and intrinsic curvatures can be found in Bates and Watts (1980).

The statistical significance of bold upper T Superscript tau and bold upper T Superscript eta and the corresponding RMS values can be assessed by comparing these values with 1 slash StartRoot upper F EndRoot, where F is the upper alpha times 100 percent-sign quantile of an F distribution with p and n minus p degrees of freedom (Bates and Watts 1980).

One motivation for fitting a nonlinear model in a different parameterization is to obtain a particular interpretation and to give parameter estimators more close-to-linear behavior. Example 88.4 shows how changing the parameterization of a four-parameter logistic model can reduce the parameter-effects curvature and can yield a useful parameter interpretation at the same time. In addition, Example 88.6 shows a nonlinear model with a high intrinsic curvature and the corresponding diagnostics.

Leverage in Nonlinear Regression

In contrast to linear regression, there are several measures of leverage in nonlinear regression. Furthermore, in nonlinear regression, the effect of a change in the ith response on the ith predicted value might depend on both the size of the change and the ith response itself (St. Laurent and Cook 1992). As a result, some observations might show superleverage —namely, leverages in excess of one (St. Laurent and Cook 1992).

PROC NLIN provides two measures of leverages: tangential and Jacobian leverages through the PLOTS option in the PROC NLIN statement and the H= and J= options of OUTPUT statement. Tangential leverage, bold upper H Subscript i, is based on approximating the nonlinear model with a linear model that parameterizes the tangent plane at the least squares parameter estimators. In contrast, Jacobian leverage, bold upper J Subscript i, is simply defined as the instantaneous rate of change in the ith predicted value with respect to the ith response (St. Laurent and Cook 1992).

The mathematical formulas for tangential and Jacobian leverages are

StartLayout 1st Row 1st Column bold upper H Subscript i 2nd Column equals w Subscript i Baseline bold x Subscript i Baseline left-parenthesis bold upper X prime bold upper W bold upper X right-parenthesis Superscript negative 1 Baseline bold x Subscript i Superscript prime Baseline 2nd Row 1st Column bold upper J Subscript i 2nd Column equals w Subscript i Baseline bold x Subscript i Baseline left-parenthesis bold upper X prime bold upper W bold upper X minus left-bracket bold upper W bold e right-bracket left-bracket bold upper H right-bracket right-parenthesis Superscript negative 1 Baseline bold x Subscript i Superscript prime Baseline comma EndLayout

where bold e is the vector of residuals, bold upper W is the diagonal weight matrix if you specify the special variable _WEIGHT_ and otherwise the identity matrix, i indexes the corresponding quantities for the ith observation, and the bold upper X matrix is defined in the section Notation for Nonlinear Regression Models. The brackets left-bracket period right-bracket left-bracket period right-bracket indicate column multiplication as defined in Bates and Watts (1980). The preceding formula for tangential leverage holds if the gradient, Marquardt, or Gauss methods are used. For the Newton method, the tangential leverage is set equal to the Jacobian leverage.

In a model with a large intrinsic curvature, the Jacobian and tangential leverages can be very different. In fact, the two leverages are identical only if the model provides an exact fit to the data (bold e equals 0) or the model is intrinsically linear (St. Laurent and Cook 1993). This is also illustrated by the leverage plot and nonlinearity measures provided in Example 88.6.

Local Influence in Nonlinear Regression

St. Laurent and Cook (1993) suggest using l Subscript max, the direction that yields the maximum normal curvature, to assess the local influence of an additive perturbation to the response variable on the estimation of the parameters and variance of a nonlinear model. Defining the normal curvature components

StartLayout 1st Row 1st Column bold upper C Subscript beta 2nd Column equals max Subscript l Baseline StartFraction 2 Over sigma squared EndFraction l prime bold upper J l 2nd Row 1st Column bold upper C Subscript sigma 2nd Column equals max Subscript l Baseline StartFraction 4 Over sigma squared EndFraction l prime bold upper P Subscript e Baseline l EndLayout

where bold upper J is the Jacobian leverage matrix and bold upper P Subscript e Baseline equals bold e bold e Superscript prime Baseline slash left-parenthesis bold e prime bold e right-parenthesis, you choose the l Subscript max that results in the maximum of the two curvature components (St. Laurent and Cook 1993). PROC NLIN provides l Subscript max through the PLOTS option in the PROC NLIN statement and the LMAX= option in the OUTPUT statement. Example 88.6 shows a plot of l Subscript max for a model with high intrinsic curvature.

Residuals in Nonlinear Regression

If a nonlinear model is intrinsically nonlinear, using the residuals bold e equals bold y minus ModifyingAbove bold y With bold caret for diagnostics can be misleading (Cook and Tsai 1985). This is due to the fact that in correctly specified intrinsically nonlinear models, the residuals have nonzero means and different variances, even when the original error terms have identical variances. Furthermore, the covariance between the residuals and the predicted values tends to be negative semidefinite, complicating the interpretation of plots based on bold e (Cook and Tsai 1985).

Projected residuals are proposed by Cook and Tsai (1985) to overcome these shortcomings of residuals, which are henceforth called raw (ordinary) residuals to differentiate them from their projected counterparts. Projected residuals have zero means and are uncorrelated with the predicted values. In fact, projected residuals are identical to the raw residuals in intrinsically linear models.

PROC NLIN provides raw and projected residuals, along with their standardized forms. In addition, the mean or expectation of the raw residuals is available. These can be accessed with the PLOTS option in the PROC NLIN statement and the OUTPUT statement options PROJRES=, PROJSTUDENT=, RESEXPEC=, RESIDUAL= and STUDENT=.

Denote the projected residuals by bold e Subscript p and the expectation of the raw residuals by normal upper E left-bracket bold e right-bracket. Then

StartLayout 1st Row 1st Column bold e Subscript p 2nd Column equals left-parenthesis bold upper I Subscript n Baseline minus bold upper P Subscript x h Baseline right-parenthesis bold e 2nd Row 1st Column normal upper E left-bracket bold e Subscript i Baseline right-bracket 2nd Column equals minus StartFraction sigma squared Over 2 EndFraction sigma-summation Underscript j equals 1 Overscript n Endscripts bold upper P overTilde Subscript x comma i j Baseline trace left-parenthesis left-bracket bold upper H Subscript j Baseline right-bracket left-parenthesis bold upper X prime bold upper X right-parenthesis Superscript negative 1 Baseline right-parenthesis EndLayout

where bold e Subscript i is the ith observation raw residual, bold upper I Subscript n is an n-dimensional identity matrix, bold upper P Subscript x h is the projector onto the column space of left-parenthesis bold upper X vertical-bar bold upper H right-parenthesis, and bold upper P overTilde Subscript x Baseline equals bold upper I Subscript n Baseline minus bold upper P Subscript x. The preceding formulas are general with the projectors defined accordingly to take the weighting into consideration. In unweighted least squares, normal upper E left-bracket bold e right-bracket reduces to

StartLayout 1st Row 1st Column normal upper E left-bracket bold e right-bracket 2nd Column equals minus one-half sigma squared bold upper Q overTilde bold a EndLayout

with bold upper Q overTilde being the last n minus p columns of the bold upper Q matrix in the QR decomposition of bold upper X and the left-parenthesis n minus p right-parenthesis dimensional vector bold a being defined in terms of the intrinsic acceleration array

StartLayout 1st Row 1st Column bold a Subscript i 2nd Column equals sigma-summation Underscript j equals 1 Overscript p Endscripts left-bracket bold upper A Subscript i plus p Baseline right-bracket Subscript j j EndLayout

Standardization of the projected residuals requires the variance of the projected residuals. This is estimated using the formula (Cook and Tsai 1985)

StartLayout 1st Row 1st Column sigma Subscript p Superscript 2 2nd Column equals StartFraction bold e prime Subscript p Baseline bold e Subscript p Baseline Over trace left-parenthesis bold upper I Subscript n Baseline minus bold upper P Subscript x h Baseline right-parenthesis EndFraction EndLayout

The standardized raw and projected residuals, denoted by bold e overTilde and bold e overTilde Subscript p respectively, are given by

StartLayout 1st Row 1st Column bold e overTilde 2nd Column equals StartFraction StartRoot w Subscript i Baseline EndRoot bold e Over sigma StartRoot 1 minus bold upper P Subscript x comma i i Baseline EndRoot EndFraction 2nd Row 1st Column bold e overTilde Subscript p 2nd Column equals StartFraction StartRoot w Subscript i Baseline EndRoot bold e Subscript p Baseline Over sigma StartRoot 1 minus bold upper P Subscript x h comma i i Baseline EndRoot EndFraction EndLayout

The use of raw and projected residuals for diagnostics in nonlinear regression is illustrated in Example 88.6.

Profiling Parameters and Assessing the Influence of Observations on Parameter Estimates

The global measures of nonlinearity, discussed in the preceding section, are very useful for assessing the overall nonlinearity of the model. However, the impact of global nonlinearity on inference regarding subsets of the parameter set cannot be easily determined (Cook and Tsai 1985). The impact of the nonlinearity on the uncertainty of individual parameters can be efficiently described by profile t plots and confidence curves (Bates and Watts 1988; Cook and Weisberg 1990).

A profile t plot for parameter beta is a plot of the likelihood ratio pivotal statistic, tau left-parenthesis beta right-parenthesis, and the corresponding Wald pivotal statistic, sigma left-parenthesis beta right-parenthesis (Bates and Watts 1988). The likelihood ratio pivotal statistic is defined as

tau left-parenthesis beta right-parenthesis equals sign left-parenthesis beta minus ModifyingAbove beta With caret right-parenthesis upper L left-parenthesis beta right-parenthesis

with

upper L left-parenthesis beta right-parenthesis equals left-parenthesis StartFraction SSE left-parenthesis beta comma normal upper Theta overTilde right-parenthesis minus SSE left-parenthesis ModifyingAbove beta With caret comma ModifyingAbove normal upper Theta With caret right-parenthesis Over mse EndFraction right-parenthesis Superscript 1 slash 2

where beta is the profile parameter and normal upper Theta refers to the remaining parameters. normal upper S normal upper S normal upper E left-parenthesis beta comma normal upper Theta overTilde right-parenthesis is the sum of square errors where the profile parameter beta is constrained at a given value and normal upper Theta overTilde is the least squares estimate of normal upper Theta conditional on a given value of beta. In contrast, normal upper S normal upper S normal upper E left-parenthesis ModifyingAbove beta With caret comma ModifyingAbove normal upper Theta With caret right-parenthesis is the sum of square errors for the full model. For linear models, tau left-parenthesis beta right-parenthesis matches sigma left-parenthesis beta right-parenthesis, which is defined as

sigma left-parenthesis beta right-parenthesis equals StartFraction beta minus ModifyingAbove beta With caret Over stderr Subscript beta Baseline EndFraction

where beta is the constrained value, ModifyingAbove beta With caret is the estimated value for the parameter, and stderr Subscript beta is the standard error for the parameter. Usually a profile t plot is overlaid with a reference line that passes through the origin and has a slope of one. PROC NLIN follows this convention.

A confidence curve for a particular parameter is useful for validating Wald-based confidence intervals for the parameter against likelihood-based confidence intervals (Cook and Weisberg 1990). A confidence curve contains a scatter plot of the constrained parameter value versus normal upper L left-parenthesis beta right-parenthesis. The Wald-based confidence intervals are overlaid as two straight lines that pass through left-parenthesis 0 comma ModifyingAbove beta With caret right-parenthesis with a slope of plus-or-minus stderr Subscript beta. Hence, for different levels of significance, you can easily compare the Wald-based confidence intervals against the corresponding confidence intervals that are based on the likelihood ratio. Cook and Weisberg (1990) recommend that you report a single Wald-based confidence interval only if there is a good agreement between the Wald and likelihood intervals up to at least the 99% confidence level.

Compared to local influence, the leave-one-out method performs a more complete analysis of the influence of observations on the values of parameter estimates. In this method, jackknife resampling removes each observation in turn and fits the model to the remaining data set. Hence, a data set with n observations will have n corresponding data sets with n–1 observations. The impact of each observation on a parameter is assessed by the absolute relative percentage change in the value of the parameter compared with the reference value from the full data.

Bootstrap Resampling and Estimation

Bootstrap resampling and estimation methods can be used to produce confidence intervals and covariance matrix estimates that have second-order, script upper O left-parenthesis 1 slash n right-parenthesis accuracy, where n is the number of observations (DiCiccio and Efron 1996). In contrast, the standard Wald-based confidence interval has first-order, script upper O left-parenthesis 1 slash StartRoot n EndRoot right-parenthesis accuracy. Bootstrap methods achieve this higher accuracy at the cost of an orders-of-magnitude increase in numerical computation compared to standard asymptotic approximations. However, dramatic increases in performance and decreases in the cost of numerical computation have made bootstrap methods very attractive for routine statistical inference (MacKinnon 2002).

PROC NLIN samples as many bootstrap sample data sets (replicates) as you specify in the NSAMPLES= option and performs least squares fit on each replicate. For each least squares fit, PROC NLIN uses the original parameter estimates as starting values for the model parameters. The statistics from the least squares fits that converge are collected and used to produce confidence intervals, covariance and correlation matrices, and histograms and scatter plots of the bootstrap parameter estimates.

Each replicate is obtained by sampling the residuals instead of the input data set. The sampled residuals are used to simulate the response by using a bootstrap data generating process (DGP). PROC NLIN’s bootstrap DGP produces replicates that contain the same number of observations as the number of usable observations in the input data set. In fact, the bootstrap DGP for a particular replicate starts by discarding the input data set observations that PROC NLIN deems unusable during the original least squares estimation. The next step of the bootstrap DGP for a replicate can be represented by the formula

upper Y overTilde Subscript i Baseline equals f left-parenthesis ModifyingAbove bold-italic beta With caret semicolon bold z prime Subscript i right-parenthesis plus epsilon overTilde Subscript i

where upper Y overTilde Subscript i is the ith simulated response, ModifyingAbove bold-italic beta With caret is the original least squares parameter estimate, bold z prime Subscript i is the ith regressor vector, and epsilon overTilde Subscript i is the ith simulated error.

PROC NLIN makes several bootstrap DGP types available in the DGP= option of the BOOTSTRAP statement. These bootstrap DGP types differ only in how they obtain the ith simulated error, epsilon overTilde Subscript i. In general, epsilon overTilde Subscript i can be represented as

epsilon overTilde Subscript i Baseline equals bold s Subscript r Baseline ModifyingAbove bold e With caret Subscript r

where r is a uniform random integer between 1 and the number of usable observations, bold s Subscript r is a scale factor that depends on the chosen bootstrap DGP options, and ModifyingAbove bold e With caret Subscript r is the rth residual obtained from the rth usable observation of the original least squares fit. The scale factor, bold s Subscript r, that captures the differences among the various bootstrap DGP types takes one of the following values:

bold s Subscript r Baseline equals StartLayout Enlarged left-brace 1st Row 1st Column 1 2nd Column if DGP equals RESIDUAL left-parenthesis RAW right-parenthesis 2nd Row 1st Column StartRoot StartFraction bold n Over bold n minus bold p EndFraction EndRoot 2nd Column if DGP equals RESIDUAL left-parenthesis ADJSSE right-parenthesis 3rd Row 1st Column StartFraction 1 Over StartRoot 1 minus bold upper H Subscript r Baseline EndRoot EndFraction 2nd Column if DGP equals RESIDUAL left-parenthesis TAN right-parenthesis 4th Row 1st Column StartFraction 1 Over StartRoot 1 minus bold upper J Subscript r Baseline EndRoot EndFraction 2nd Column if DGP equals RESIDUAL left-parenthesis JAC right-parenthesis 5th Row 1st Column StartFraction gamma Over StartRoot 1 minus bold upper H Subscript r Baseline EndRoot EndFraction 2nd Column if DGP equals WILD EndLayout

In the preceding formula, bold n is the number of usable observations, bold p is the number of model parameters, and bold upper H Subscript r and bold upper J Subscript r are the rth tangential and Jacobian leverages, respectively. For the WILD bootstrap DGP (Wu 1986), which is the only bootstrap DGP type exclusively available for weighted least squares, gamma is a random number given by

gamma equals StartLayout Enlarged left-brace 1st Row 1st Column minus StartFraction StartRoot 5 EndRoot minus 1 Over 2 EndFraction 2nd Column normal w normal i normal t normal h normal p normal r normal o normal b normal a normal b normal i normal l normal i normal t normal y StartFraction StartRoot 5 EndRoot plus 1 Over 2 StartRoot 5 EndRoot EndFraction 2nd Row 1st Column StartFraction StartRoot 5 EndRoot plus 1 Over 2 EndFraction 2nd Column normal w normal i normal t normal h normal p normal r normal o normal b normal a normal b normal i normal l normal i normal t normal y StartFraction StartRoot 5 EndRoot minus 1 Over 2 StartRoot 5 EndRoot EndFraction EndLayout

PROC NLIN makes three types of bootstrap confidence intervals available in the BOOTCI option in the BOOTSTRAP statement. These confidence intervals are the percentile, normal, and bias-corrected bootstrap confidence intervals. The computational details of these confidence intervals for beta Subscript i, the ith model parameter, follow. For simplicity of notation, denote beta Subscript i as beta. Also, let B represent the number of replicates for which the least squares fit converges

The option that computes the percentile bootstrap confidence interval, BOOTCI(PERC), does so by computing the 100 left-parenthesis alpha slash 2 right-parenthesisth and 100 left-parenthesis 1 minus alpha slash 2 right-parenthesisth percentiles from the bootstrap parameter estimates. alpha is from the ALPHA= option in the PROC NLIN statement. These percentiles are computed as follows. Let beta overTilde Subscript 1 Baseline comma beta overTilde Subscript 2 Baseline commacomma beta overTilde Subscript upper B Baseline represent the ordered values of the bootstrap estimates for beta. Let the kth weighted average percentile be y, set p equals StartFraction k Over 100 EndFraction, and let

n p equals j plus g

where j is the integer part of n p and g is the fractional part of n p. Then the kth percentile, y, is given by

y equals StartLayout Enlarged left-brace 1st Row 1st Column one-half left-parenthesis beta overTilde Subscript j Baseline plus beta overTilde Subscript j plus 1 Baseline right-parenthesis 2nd Column normal i normal f g equals 0 2nd Row 1st Column beta overTilde Subscript j plus 1 Baseline 2nd Column normal i normal f g greater-than 0 EndLayout

which corresponds to the default percentile definition of the UNIVARIATE procedure.

In contrast, the BOOTCI(NORMAL) option in the BOOTSTRAP statement computes the normal bootstrap confidence interval by approximating the distribution of the bootstrap parameter estimates as a normal distribution. Consequently, the normal bootstrap confidence interval, for alpha level, is given by

ModifyingAbove beta With caret plus-or-minus stdb Subscript beta overTilde times normal z Superscript left-parenthesis 1 minus alpha slash 2 right-parenthesis

where ModifyingAbove beta With caret is the original least squares parameter estimate, stdb Subscript beta overTilde is the standard deviation of the bootstrap parameter estimates, beta overTilde; and normal z Superscript left-parenthesis 1 minus alpha slash 2 right-parenthesis is the 100 left-parenthesis 1 minus alpha slash 2 right-parenthesisth percentile of the standard normal distribution.

The BOOTSTRAP statement option that computes the bias-corrected bootstrap confidence interval, BOOTCI(BC), does so by making use of the cumulative distribution function (CDF), upper G left-parenthesis beta overTilde right-parenthesis, of the bootstrap parameter estimates to correct for the upper and lower endpoints of the alpha level. The bias-corrected bootstrap confidence interval is given by

upper G Superscript negative 1 Baseline left-parenthesis normal upper Phi left-parenthesis 2 z 0 plus-or-minus z Subscript alpha Baseline right-parenthesis right-parenthesis

where normal upper Phi is the standard normal CDF, z Subscript alpha Baseline equals normal upper Phi Superscript negative 1 Baseline left-parenthesis alpha right-parenthesis, and z 0 is a bias correction given by

z 0 equals normal upper Phi Superscript negative 1 Baseline left-parenthesis StartFraction normal upper N left-parenthesis beta overTilde less-than-or-equal-to ModifyingAbove beta With caret right-parenthesis Over normal upper B EndFraction right-parenthesis

where ModifyingAbove beta With caret is the estimate of beta from the original least squares fit and normal upper N left-parenthesis beta overTilde less-than-or-equal-to ModifyingAbove beta With caret right-parenthesis is the number of bootstrap estimates, beta overTilde, that are less than or equal to ModifyingAbove beta With caret.

In addition, PROC NLIN produces bootstrap estimates of the covariance and correlation matrices. For the ith and jth model parameters, the covariance is estimated by

Cov left-parenthesis beta overTilde Subscript i Baseline comma beta overTilde Subscript j Baseline right-parenthesis equals sigma-summation Underscript k equals 1 Overscript k equals normal upper B Endscripts StartFraction left-parenthesis beta overTilde Subscript i k Baseline minus beta overTilde overbar Subscript i Baseline right-parenthesis left-parenthesis beta overTilde Subscript j k Baseline minus beta overTilde overbar Subscript j Baseline right-parenthesis Over normal upper B minus 1 EndFraction

where the sum runs over the nonmissing bootstrap parameter estimates. The bootstrap correlation matrix is estimated by scaling the bootstrap covariance matrix

Corr left-parenthesis beta overTilde Subscript i Baseline comma beta overTilde Subscript j Baseline right-parenthesis equals StartFraction Cov left-parenthesis beta overTilde Subscript i Baseline comma beta overTilde Subscript j Baseline right-parenthesis Over normal s normal t normal d normal b Subscript beta overTilde Sub Subscript i Subscript Baseline normal s normal t normal d normal b Subscript beta overTilde Sub Subscript j Subscript Baseline EndFraction

where stdb Subscript beta overTilde Sub Subscript i and stdb Subscript beta overTilde Sub Subscript j are the bootstrap standard deviations for the ith and jth parameters.

Last updated: December 09, 2022