DOC PREVIEW
LSU EXST 7034 - Multiple Regression

This preview shows page 1 out of 2 pages.

Save
View full document
View full document
Premium Document
Do you want full access? Go Premium and unlock all 2 pages.
Access to all documents
Download any document
Ad free experience
Premium Document
Do you want full access? Go Premium and unlock all 2 pages.
Access to all documents
Download any document
Ad free experience

Unformatted text preview:

EXST7034 : Regression Techniques Geaghan Multiple Regression Diagnostic variables and criteria Page 1 08-decision criteria.doc Criteria for the interpretation of selected statistics from the SAS output A) General regression diagnostics 1) Adjusted R2 : RnnpSSErrorSSTotalnnpRadj221111=−−−FHGIKJ=−−FHGIKJ−bgbgch a) This is intended to be an adjustment to R2 for additional variables in the model b) Unlike the usual R2, this value can decrease as more variables are entered in the model if the variables do not account for sufficient additional variation (equal to the MSE). 2) Standardized regression coefficient bj' . bj' = bj (Sxj / Sy) a) Unlike the usual regression coefficient, the magnitude of the standardized coefficient provides a meaningful comparison among the regression coefficients. b) Larger standardized regression coefficients have more impact on the calculation of the predicted value and are more "important". 3) Partial correlations a) Squared semi-partial correlation TYPE I = SCORR1 = SeqSSXj / SSTotal b) Squared partial correlation TYPE I = PCORR1 = SeqSSXj / (SeqSSXj + SSError*) c) Squared semi-partial correlation TYPE II = SCORR2 = PartialSSXj / SSTotal d) Squared partial correlation TYPE II = PCORR2 = PartialSSXj / (PartialSSXj + SSError) e) Note that for regression, TYPE II SS and TYPE III SS are the same. B) Residual Diagnostics 1) The hat matrix main diagonal elements, hii (Hat Diag , H values in SAS) , called "leverage values", they are used to detect outliers in X space. . This can also identify substantial extrapolation of new values. As a general rule, hii values greater than 0.5 are "large" while those between 0.2 and 0.5 are moderately large. Also look for a leverage value which is noticeably larger than the next largest, leaving a gap between values. a) The hii values sum to p mean,⎯hii = p/n (note that this is < 1) b) A value may be an outlier if it is more than twice the value⎯hii (i.e. hii > 2p/n). 2) Studentized residuals (“Student Residual” in SAS). a) There are two versions: Simpler calculation = ei / root(MSE) (the “semistudentized” residual) More common application = ei / root(MSE * (1-hii)) [SAS produces these] b) We already assume these are normally distributed, so these values would approximately follow a t distribution, where for large samples about 65% are between -1 and +1 about 95% are between -2 and +2 about 99% are between -2.6 and +2.6 3) Deleted Studentized residuals (“RStudent” in SAS). Also called externally studentized residual. a) There are also two versions as with the studentized residuals above Deleted Semistudentized residual = ei(i) / root(MSE(i)) Deleted Studentized residual = ei(i) / root(MSE(i) *1-hii) [SAS produces these values] b) As with the studentized residuals above these values would approximately follow a t distribution. The text recommends a Bonferroni adjustment c) These are one of the best indicators of outliers, but they can only detect one outlier at a time. C) Influence Diagnostics 1) ()()ˆˆiiiiiiiYYDFFITSMSE h−= , measures the difference in fits judged by the change in predicted value when the point is omitted. a) This is a standardized value and can be interpreted as the number of standard deviation unitsEXST7034 : Regression Techniques Geaghan Multiple Regression Diagnostic variables and criteria Page 2 08-decision criteria.doc b) for small to medium size databases, DFFITS should not exceed 1, while for large databases it should not exceed 2pn 2) ()()iiiikkbbDFBETASMSE c−= where ckk is from the X´X-1. DFBETAS measures the difference in fits judged by the change in the values of the regression coefficients a) note that this is also a standardized value b) for small to medium size databases, DFBETAS should not exceed 1, while for large databases it should not exceed 2/sqrt(n) 3) 2''()21()()2ˆˆ()()()' (1 )njjijjjiiiiiYYbb XXbbehCook s DpMSE pMSE h pMSE=−−−⎡⎤== =⎢⎥−⎣⎦∑ (D is for distance) a) derived from a concept of a joint confidence region for all p regression coefficients b) this does not follow an F distribution, but it is useful to compare it to the percentiles of the F distribution [F1-α; p, n-p] where a change of < 10th or 20th percentile shows little effect, while the 50th percentile is considered large D) multicollinearity Diagnostics 1) VIF is related to the severity of multicolinearity (VIF = (1–Rk2)-1 a) a standardized estimate of regression coefficients would be expected to have a value of 1 if the regressors are uncorrelated b) If the mean of this value is much greater than 1, serious problems are indicated. c) No single VIF should exceed 10 2) Tolerance is the inverse of VIF, where Tolerancek = 1–Rk2 3) The Condition number (a multivariate evaluation) a) Eigenvalues are extracted from the regressors, These are variances of linear combinations of the regressors, and go from larger to smaller. b) If one or more are zero (at the end) then the matrix is not full rank. c) These sum to p, and if the Xk are independent, each would equal 1 d) The condition number is the square root of the ratio of the largest (always the first) to each of the others. e) If this value exceeds 30 then multicollinearity may be a problem. E) Model Evaluation and Validation 1) R2p, AdjR2p and MSEp can be used to graphically compare and evaluate models. The subscript p refers to the number of parameters in the model 2) Mallow's Cp criterion a) Use of this statistic presumes no bias in the full model MSE, so the full model should be carefully chosen to have little or no multicollinearity b) Cp criterion = (SSEp / TrueMSE) – (n – 2p) c) The Cp statistics will be approximately equal to p if there is no bias in the regression model 3) PRESSp criterion (PRESS = Prediction SS) 2()1ˆ()npiiiiPRESS Y Y==−∑ a) This criterion is based on deleted residuals. b) There are n deleted residuals in each regression, and PRESSp is the SS of deleted residuals c) This value should approximately equal the MSE if predictions are good, it will get larger as predictions are poorer d) They may be plotted, and the smaller PRESS statistic models represent better predictive models. This statistics can also be used for model validation and model


View Full Document
Download Multiple Regression
Our administrator received your request to download this document. We will send you the file to your email shortly.
Loading Unlocking...
Login

Join to view Multiple Regression and access 3M+ class-specific study document.

or
We will never post anything without your permission.
Don't have an account?
Sign Up

Join to view Multiple Regression 2 2 and access 3M+ class-specific study document.

or

By creating an account you agree to our Privacy Policy and Terms Of Use

Already a member?