The NLIN Procedure

Notation for Nonlinear Regression Models

This section briefly introduces the basic notation for nonlinear regression models that applies in this chapter. Additional notation is introduced throughout as needed.

The left-parenthesis n times 1 right-parenthesis vector of observed responses is denoted as bold y. This vector is the realization of an left-parenthesis n times 1 right-parenthesis random vector bold upper Y. The NLIN procedure assumes that the variance matrix of this random vector is sigma squared bold upper I. In other words, the observations have equal variance (are homoscedastic) and are uncorrelated. By defining the special variable _WEIGHT_ in your NLIN programming statements, you can introduce heterogeneous variances. If a _WEIGHT_ variable is present, then normal upper V normal a normal r left-bracket bold upper Y right-bracket equals sigma squared bold upper W Superscript negative 1, where bold upper W is a diagonal matrix containing the values of the _WEIGHT_ variable.

The mean of the random vector is represented by a nonlinear model that depends on parameters beta 1 comma ellipsis comma beta Subscript p Baseline and regressor (independent) variables z 1 comma ellipsis comma z Subscript k Baseline:

normal upper E left-bracket upper Y Subscript i Baseline right-bracket equals f left-parenthesis beta 1 comma beta 2 comma ellipsis comma beta Subscript p Baseline semicolon z Subscript i Baseline 1 Baseline comma ellipsis comma z Subscript i k Baseline right-parenthesis

In contrast to linear models, the number of regressor variables (k) does not necessarily equal the number of parameters (p) in the mean function f left-parenthesis right-parenthesis. For example, the model fitted in the next subsection contains a single regressor and two parameters.

To represent the mean of the vector of observations, boldface notation is used in an obvious extension of the previous equation:

normal upper E left-bracket bold upper Y right-bracket equals bold f left-parenthesis bold-italic beta semicolon bold z 1 comma ellipsis comma bold z Subscript k Baseline right-parenthesis

The vector bold z 1, for example, is an left-parenthesis n times 1 right-parenthesis vector of the values for the first regressor variables. The explicit dependence of the mean function on bold-italic beta and/or the bold z vectors is often omitted for brevity.

In summary, the stochastic structure of models fit with the NLIN procedure is mathematically captured by

StartLayout 1st Row 1st Column bold upper Y 2nd Column equals bold f left-parenthesis bold-italic beta semicolon bold z 1 comma ellipsis comma bold z Subscript k Baseline right-parenthesis plus bold-italic epsilon 2nd Row 1st Column normal upper E left-bracket bold-italic epsilon right-bracket 2nd Column equals bold 0 3rd Row 1st Column normal upper V normal a normal r left-bracket bold-italic epsilon right-bracket 2nd Column equals sigma squared bold upper I EndLayout

Note that the residual variance sigma squared is typically also unknown. Since it is not estimated in the same fashion as the other p parameters, it is often not counted in the number of parameters of the nonlinear regression. An estimate of sigma squared is obtained after the model fit by the method of moments based on the residual sum of squares.

A matrix that plays an important role in fitting nonlinear regression models is the left-parenthesis n times p right-parenthesis matrix of the first partial derivatives of the mean function bold f with respect to the p model parameters. It is frequently denoted as

bold upper X equals StartFraction partial-differential bold f left-parenthesis bold-italic beta semicolon bold z 1 comma ellipsis comma bold z Subscript k Baseline right-parenthesis Over partial-differential bold-italic beta EndFraction

The use of the symbol bold upper X—common in linear statistical modeling—is no accident here. The first derivative matrix plays a similar role in nonlinear regression to that of the bold upper X matrix in a linear model. For example, the asymptotic variance of the nonlinear least-squares estimators is proportional to left-parenthesis bold upper X prime bold upper X right-parenthesis Superscript negative 1, and projection-type matrices in nonlinear regressions are based on bold upper X left-parenthesis bold upper X prime bold upper X right-parenthesis Superscript negative 1 Baseline bold upper X prime. Also, fitting a nonlinear regression model can be cast as an iterative process where a nonlinear model is approximated by a series of linear models in which the derivative matrix is the regressor matrix. An important difference between linear and nonlinear models is that the derivatives in a linear model do not depend on any parameters (see previous subsection). In contrast, the derivative matrix partial-differential bold f left-parenthesis bold-italic beta right-parenthesis slash partial-differential bold-italic beta is a function of at least one element of bold-italic beta. It is this dependence that lies at the core of the fact that estimating the parameters in a nonlinear model cannot be accomplished in closed form, but it is an iterative process that commences with user-supplied starting values and attempts to continually improve on the parameter estimates.

Last updated: December 09, 2022