In this section, degrees of freedom (df) refers to the denominator degrees of freedom for F statistics in hypothesis testing. It also refers to the degrees of freedom in t tests for parameter estimates and odds ratio estimates, and for computing t distribution percentiles for confidence limits of these estimates. The value of df is determined by the design degrees of freedom f and by what you specify in the DF= option in the MODEL statement.
The default df is determined as
where f is the design degrees of freedom and r is the rank of the contrast of model parameters to be tested.
The design degrees of freedom f is determined by the survey design and the variance estimation method.
For Taylor series variance estimation, the design degrees of freedom f can depend on the number of clusters, the number of strata, and the number of observations. These numbers are based on the observations that are included in the analysis; they do not count observations that are excluded from the analysis because of missing values. If all values in a stratum are excluded from the analysis as missing values, then that stratum is called an empty stratum. Empty strata are not counted in the total number of strata for the analysis. Similarly, empty clusters and missing observations are not included in the totals counts of clusters and observations that are used to compute the f for the analysis.
If you specify the MISSING option in the CLASS statement, missing values are treated as valid nonmissing levels and are included in determining the f. If you specify the NOMCAR option for Taylor series variance estimation, observations that have missing values for variables in the regression model are included. For more information about missing values, see the section Missing Values.
Using the notation that is defined in the section Notation, let be the total number of clusters if the design has a CLUSTER statement; let n be the total sample size; and let H be the number of strata if there is a STRATA statement, or 1 otherwise. Then for Taylor series variance estimation, the design degrees of freedom is
For replication variance estimation methods that include a REPWEIGHTS statement, the design degrees of freedom f equals the number of REPWEIGHTS variables, unless you specify an alternative in the DF= option in a REPWEIGHTS statement.
When the BRR method (including Fay’s method) is requested but no REPWEIGHTS statement is specified, the design degrees of freedom f equals the number of strata.
When the jackknife method or the bootstrap method is requested but no REPWEIGHTS statement is specified, the design degrees of freedom f is the same as in the Taylor series method described in the previous section.
If you do not want to use the default design degrees of freedom, then you can specify the DF=DESIGN(value) or DF=PARMADJ(value) (for Taylor method only) option in the MODEL statement, where value is a positive number. Then f=value.
However, if you specify the value in the DF= option in the MODEL statement as well as with the DF= option in a REPWEIGHTS statement, then the df is determined by the value in the MODEL statement, and the DF= option in the REPWEIGHTS statement is ignored.
If you specify DF=INFINITY in the MODEL statement, then the df is set to be infinite.
As the denominator degrees of freedom grows, an F distribution approaches a chi-square distribution, and similarly a t distribution approaches a normal distribution. Therefore, when you specify DF=INFINITY in the MODEL statement), PROC SURVEYLOGISTIC uses chi-square tests and normal distribution percentiles to construct confidence intervals.
When you use Taylor series variance estimation (by default or when you specify VARMETHOD=TAYLOR in the MODEL statement), and you are fitting a model that has many parameters relative to the design degrees of freedom, it is appropriate to modify the design degrees of freedom by using the number of nonsingular parameters p in the model (Korn and Graubard (1999, section 5.2), Rao, Scott, and Skinner (1998)). You can specify DF=PARMADJ or DF=PARMADJ(value) in the MODEL statement to request this modification only for Taylor series variance estimation method; and this option does not apply to the replication variance estimation method.
Let f be the design degrees of freedom that is described in the section Design Degrees of Freedom f for the Taylor Series Method. By default, or if you specify the DF=PARMADJ option, the df is modified as .
The global null hypothesis refers to the null hypothesis that all the explanatory effects can be eliminated and the model can contain only intercepts. By using the notations in the section Logistic Regression Models, the global null hypothesis is defined as follows:
If you have a cumulative model whose model parameters are , where
are the parameters for the intercepts and
are the parameters for the explanatory effects, then
. The number of restrictions r that are imposed on
is the number of parameters in slope parameter
:
.
If you have a generalized logit model whose model parameters are and
, then
. The number of restrictions r that are imposed on
is the total number of slope parameters in
:
.
PROC SURVEYLOGISTIC displays these tests in the "Testing Global Null Hypothesis: BETA=0" table.
For complex survey design, you can use a design-adjusted Rao-Scott likelihood ratio chi-square test to test the global null hypothesis. For information about design-adjusted chi-square tests, see Lohr (2010, Section 10.3.2), Rao and Scott (1981), Rao and Scott (1984), Rao and Scott (1987), Thomas and Rao (1987), Rao and Thomas (1989), and Thomas, Singh, and Roberts (1996).
If you specify the CHISQ(NOADJUST) option, PROC SURVEYLOGISTIC computes the likelihood ratio chi-square test without the Rao-Scott design correction. If you specify the CHISQ(FIRSTORDER) option, PROC SURVEYLOGISTIC performs a first-order Rao-Scott likelihood ratio chi-square test. If you specify the CHISQ(SECONDORDER) option, PROC SURVEYLOGISTIC performs a second-order Rao-Scott likelihood ratio chi-square test.
If you do not specify the CHISQ option, the default test depends on the design and the model. By default, PROC SURVEYLOGISTIC performs a first-order or second-order Rao-Scott likelihood ratio chi-square (Satterthwaite) test if your design is not simple random sampling or when you provide replicate weights. Otherwise, if your design is simple random sampling and you do not provide replicate weights, PROC SURVEYLOGISTIC does not make any adjustment for the likelihood ratio test. In other words:
If your design does not contain stratification nor clustering, and you do not provide replicate weights, then by default PROC SURVEYLOGISTIC performs a likelihood ratio chi-square test without any adjustment.
If your design contains either stratification or clustering, or if you provide replicate weights, then by default PROC SURVEYLOGISTIC performs a likelihood ratio chi-square test with Rao-Scott adjustment. However, the default order of the adjustment depends on the number of model parameters excluding the intercepts.
If there is more than one nonintercept parameter in the model, the default is the second-order Rao-Scott likelihood ratio test.
If there is only one nonintercept parameter in the model, there is no need to compute the second-order adjustment. Therefore, the default is the first-order Rao-Scott likelihood ratio test.
Let be the estimated parameters, let
be the estimated parameters under the global null hypothesis, and let r be the restrictions imposed on
under the global null hypothesis
. Let
be the log-likelihood function that is computed by using normalized weights.
Denote the estimated covariance matrix of under simple random sampling as
, and its partition corresponding to the r slope parameters as
. Similarly, denote the estimated covariance matrix of
under the sample design as
, and its partition corresponding to the r slope parameters as
.
Define the design effect matrix E as
Denote as the rank of E and the positive eigenvalues of the design matrix E as
.
Without the Rao-Scott design correction, the global null hypothesis is tested using either the chi-square statistics,
with r degrees of freedom, or an equivalent F statistics,
To address the impact of a complex survey design on the significance level of the likelihood ratio test, Rao and Scott (1984) proposed a first-order correction to the chi-square statistics as
where the first-order design correction,
is the average of positive eigenvalues of the design effect matrix E.
Under the null hypothesis, the first-order Rao-Scott chi-square approximately follows a chi-square distribution with
degrees of freedom.
The corresponding F statistic is
which has an F distribution with and
degrees of freedom under the null hypothesis (Thomas and Rao 1984, 1987), and df is the design degrees of freedom as described in the section Design Degrees of Freedom f for the Taylor Series Method.
Rao and Scott (1987) further proposed the second-order (Satterthwaite) Rao-Scott chi-square statistic as
where is the first-order Rao-Scott chi-square statistic and the second-order design correction is computed from the coefficient of variation of the eigenvalues of the design effect matrix E as
Under the null hypothesis, the second-order Rao-Scott chi-square approximately follows a chi-square distribution with
degrees of freedom.
The corresponding F statistic is
which has an F distribution with and
degrees of freedom under the null hypothesis (Thomas and Rao 1984, 1987), and df is the design degrees of freedom as described in the section Design Degrees of Freedom f for the Taylor Series Method.
Wald confidence intervals are sometimes called normal confidence intervals. They are based on the asymptotic normality of the parameter estimators. The % Wald confidence interval for
is given by
where is the
th percentile of the standard normal distribution,
is the pseudo-estimate of
, and
is the standard error estimate of
in the section Variance Estimation.
Linear hypotheses for can be expressed in matrix form as
where is a matrix of coefficients for the linear hypotheses and
is a vector of constants whose rank is r. The vector of regression coefficients
includes both slope parameters and intercept parameters.
Let be the MLE of
, and let
be the estimated covariance matrix that is described in the section Variance Estimation.
For the Taylor series variance estimation method, PROC SURVEYLOGISTIC computes the test statistic for the null hypothesis as
where f is the design degrees of freedom as described in the section Design Degrees of Freedom f for the Taylor Series Method.
For the replication variance estimation method, PROC SURVEYLOGISTIC computes the test statistic for the null hypothesis as
Under the ,
has an F distribution with
degrees of freedom, and the denominator degrees of freedom df is described in the section Degrees of Freedom.
As the denominator degrees of freedom grows, an F distribution approaches a chi-square distribution, and similarly a t distribution approaches a normal distribution. If you specify DF=INFINITY in the MODEL statement, PROC SURVEYLOGISTIC computes the test statistic for both Taylor series and replication methods for testing the null hypothesis as
Under ,
has an asymptotic chi-square distribution with r degrees of freedom.
For models that use less-than-full-rank parameterization (as specified by the PARAM=GLM option in the CLASS statement), a Type 3 test of an effect of interest (main effect or interaction) is a test of the Type III estimable functions that are defined for that effect. When the model contains no missing cells, performing the Type 3 test of a main effect corresponds to testing the hypothesis of equal marginal means. For more information about Type III estimable functions, see Chapter 53, The GLM Procedure, and Chapter 16, The Four Types of Estimable Functions. Also see Littell, Freund, and Spector (1991).
For models that use full-rank parameterization, all parameters are estimable when there are no missing cells, so it is unnecessary to define estimable functions. The standard test of an effect of interest in this case is the joint test that the values of the parameters associated with that effect are zero. For a model that uses effects parameterization (as specified by the PARAM=EFFECT option in the CLASS statement), performing the joint test for a main effect is equivalent to testing the equality of marginal means. For a model that uses reference parameterization (as specified by the PARAM=REF option in the CLASS statement), performing the joint test is equivalent to testing the equality of cell means at the reference level of the other model effects. For more information about the coding scheme and the associated interpretation of results, see Muller and Fetterman (2002, Chapter 14).
If there is no interaction term, the Type 3 test of an effect for a model that uses GLM parameterization is the same as the joint test of the effect for the model that uses full-rank parameterization. In this situation, the joint test is also called the Type 3 test. For a model that contains an interaction term and no missing cells, the Type 3 test of a component main effect under GLM parameterization is the same as the joint test of the component main effect under effect parameterization. Both test the equality of cell means. But this Type 3 test differs from the joint test under reference parameterization, which tests the equality of cell means at the reference level of the other component main effect. If some cells are missing, you can obtain meaningful tests only by testing a Type III estimable function, so in this case you should use GLM parameterization.
The results of a Type 3 test or a joint test do not depend on the order in which you specify the terms in the MODEL statement.
Consider a dichotomous response variable with outcomes event and nonevent. Let a dichotomous risk factor variable X take the value 1 if the risk factor is present and 0 if the risk factor is absent. According to the logistic model, the log odds function, , is given by
The odds ratio is defined as the ratio of the odds for those with the risk factor (X = 1) to the odds for those without the risk factor (X = 0). The log of the odds ratio is given by
The parameter, , associated with X represents the change in the log odds
from X = 0 to X = 1. So the odds ratio is obtained by simply exponentiating the value of the parameter associated with the risk factor. The odds ratio indicates how the odds of event change as you change X from 0 to 1. For instance,
means that the odds of an event when X = 1 are twice the odds of an event when X = 0.
Suppose the values of the dichotomous risk factor are coded as constants a and b instead of 0 and 1. The odds when become
, and the odds when
become
. The odds ratio corresponding to an increase in X from a to b is
Note that for any a and b such that . So the odds ratio can be interpreted as the change in the odds for any increase of one unit in the corresponding risk factor. However, the change in odds for some amount other than one unit is often of greater interest. For example, a change of one pound in body weight might be too small to be considered important, while a change of 10 pounds might be more meaningful. The odds ratio for a change in X from a to b is estimated by raising the odds ratio estimate for a unit change in X to the power of
, as shown previously.
For a polytomous risk factor, the computation of odds ratios depends on how the risk factor is parameterized. For illustration, suppose that Race is a risk factor with four categories: White, Black, Hispanic, and Other.
For the effect parameterization scheme (PARAM=EFFECT) with White as the reference group, the design variables for Race are as follows.
The log odds for Black is
The log odds for White is
Therefore, the log odds ratio of Black versus White becomes
For the reference cell parameterization scheme (PARAM=REF) with White as the reference cell, the design variables for race are as follows.
The log odds ratio of Black versus White is given by
For the GLM parameterization scheme (PARAM=GLM), the design variables are as follows.
The log odds ratio of Black versus White is
Consider the hypothetical example of heart disease among race in Hosmer and Lemeshow (2000, p. 51). The entries in the following contingency table represent counts.
| Race | ||||
|---|---|---|---|---|
| Disease Status | White | Black | Hispanic | Other |
| Present | 5 | 20 | 15 | 10 |
| Absent | 20 | 10 | 10 | 10 |
The computation of odds ratio of Black versus White for various parameterization schemes is shown in Table 9.
Table 9: Odds Ratio of Heart Disease Comparing Black to White
Since the log odds ratio () is a linear function of the parameters, the Wald confidence interval for
can be derived from the parameter estimates and the estimated covariance matrix. Confidence intervals for the odds ratios are obtained by exponentiating the corresponding confidence intervals for the log odd ratios. In the displayed output of PROC SURVEYLOGISTIC, the "Odds Ratio Estimates" table contains the odds ratio estimates and the corresponding t or Wald confidence intervals computed by using the covariance matrix in the section Variance Estimation. For continuous explanatory variables, these odds ratios correspond to a unit increase in the risk factors.
To customize odds ratios for specific units of change for a continuous risk factor, you can use the UNITS statement to specify a list of relevant units for each explanatory variable in the model. Estimates of these customized odds ratios are given in a separate table. Let be a confidence interval for
. The corresponding lower and upper confidence limits for the customized odds ratio
are
and
, respectively, (for
); or
and
, respectively, (for c < 0). You use the CLODDS option in the MODEL statement to request confidence intervals for the odds ratios.
For a generalized logit model, odds ratios are computed similarly, except D odds ratios are computed for each effect, corresponding to the D logits in the model.
The predicted mean score of an observation is the sum of the ordered values (shown in the "Response Profile" table) minus one, weighted by the corresponding predicted probabilities for that observation; that is, the predicted means score is , where D + 1 is the number of response levels and
is the predicted probability of the dth (ordered) response.
A pair of observations with different observed responses is said to be concordant if the observation with the lower-ordered response value has a lower predicted mean score than the observation with the higher-ordered response value. If the observation with the lower-ordered response value has a higher predicted mean score than the observation with the higher-ordered response value, then the pair is discordant. If the pair is neither concordant nor discordant, it is a tie. Enumeration of the total numbers of concordant and discordant pairs is carried out by categorizing the predicted mean score into intervals of length and accumulating the corresponding frequencies of observations.
Let N be the sum of observation frequencies in the data. Suppose there are a total of t pairs with different responses, of them are concordant,
of them are discordant, and
of them are tied. PROC SURVEYLOGISTIC computes the following four indices of rank correlation for assessing the predictive ability of a model:
Note that c also gives an estimate of the area under the receiver operating characteristic (ROC) curve when the response is binary (Hanley and McNeil 1982).
For binary responses, the predicted mean score is equal to the predicted probability for Ordered Value 2.