Saturday, November 2, 2019

Multiple Regression Essay Example | Topics and Well Written Essays - 1500 words

Multiple Regression - Essay Example We use a State-wide data set that includes a record of property crimes rates (CRIMES) as well as a record on per capita income (PINCOME), school dropout rates (DROPOUT), precipitation amounts (PRECIP), percentage of public aid recipients (PUBAID), population density (DENSITY), public aid for families with kids in terms of dollars received (KIDS), percentage of unemployed workers (UNEMPLOY), percentage of population living in urban areas. The methodology that we use is that of multiple regression analysis to obtain the magnitude and signs of the coefficients and t and F-tests obtain whether the respective coefficients are significant, individually, or jointly. The regression equation we estimate is the following: (1) Results In this section we present the results of the analysis. Table 1 presents the results of the estimation of equation (1). Table 1: Results of simple OLS regression, all variables included Before interpreting the coefficients we look at the individual and joint signi ficances of the estimated coefficients. From the upper right hand panel we find that F(8, 41) = 11.43 and prob>F =0.000. Recall that the null hypothesis of the f-test is that all coefficients are jointly zero. From the computed statistic and the associated p-value we reject the null hypothesis. Thus, at least one of the parameters is non-zero. Now, from the lower panel in table 1 looking at the 3rd column (t-values) and the 4th column (p values) we can identify which coefficients are significantly different from zero. The null hypothesis of the t-test is that the coefficient in question is equal to zero while the alternative hypothesis is that it is non-zero. Recall that the 5% critical value for the two-sided t-test is 1.96. Looking at the elements from column 3 and 4, we find that only the variables DROPOUT, DENSITY and and URBAN are associated with coefficients that are statistically significantly different from zero. We fail to find evidence in the data that suggests that the nu ll hypothesis is false for the rest of the coefficients as well as the intercept. The upper right hand corner also presents the R-squared and the adjusted R-squared values which show that the fit is decent (anything greater than 50% on these scales is considered decent). In table 2 below, we take a look at whether our results may have been affected by the presence of multi-collnearity. The t-tests of significance showed that there were two significant coefficients and this was further confirmed by the f-test. If multicollinearity is present to any worrisome degree, typically, although the f-test rejects the null, we fail to identify any statistically significant coefficients in the t-test. Thus, there is little evidence of multi-collinearity. Additionally, we also look at the VIFs in table 2 below. Observe that the mean VIF is only 2.46 and the VIF for PINCOME is a bit large (although significantly smaller than 10, the standard indicator of severe and worrisome multicollinearity). T able 2: Testing for presence of multi-colinearity, VIFs Now, we rerun the regression incorporating only the significant variables in the specification. The model is specified as follows: The results of this regression are presented in table 3. Table 3: Results of OLS estimation including only the significant variables Observe first from the table above that the signs and the significances of the included variables remain the same. That is

No comments:

Post a Comment

Note: Only a member of this blog may post a comment.