Question

In: Statistics and Probability

A MLR model have LIFE (y) as the response variable, and MALE (x1), BIRTH (x2), DIVO...

A MLR model have LIFE (y) as the response variable, and MALE (x1), BIRTH (x2), DIVO (x3), BEDS (x4), EDUC (x5), and INCO (x6), as predictors.

I know you can use first fit the model using lm(y~x) then use anova(model) to check the SSreg,my question is, what is the difference between  SSreg(β2|β0,β3) and SSreg(β1|β0,β3,β2)? What should you put as the argument of lm() function with respect to (β2|β0,β3) and (β1|β0,β3,β2)

Solutions

Expert Solution

SSreg means that sum squares of the regression model. Elaborately, it means how much variability has been explained by the proposed model. The more the SSreg of a model is, the more variability has been explained by the model. SSreg(β2|β0,β3), SSreg(β1|β0,β3,β2) - this notation is called the extra sum of square. For an MLR model the greater the number of covariates, the greater variability is being explained and hence the bigger the value of the SSreg is for the bigger model as compared to the smaller one. Now, the extra sum of square is the additional sum of squares of regression which is obtained when adding a covariate in the model where already some of the covariates are already there.

Here "SSreg(β2|β0,β3)" defines that extra sum of squares of regression obtained by adding x2 as a covariate in the model where x0 and x3 were already present. Mathematically, SSreg(β2|β0,β3) = sum of squares of regression when x0, x3, x2 are the covariates - sum of squares of regression when x0, x3 are the covariates.

Similarly, "SSreg(β1|β0,β3,β2)" defines that extra sum of squares of regression obtained by adding x1 as a covariate in the model where x0, x3, and x2 were already present. Mathematically, SSreg(β1|β0,β3,β2) = sum of squares of regression when x0, x3, x2, x1 are the covariates - sum of squares of regression when x0, x3, x2 are the covariates.

Hence the difference is that, while SSreg(β2|β0,β3) means that the extra variability explained due to the addition of the covariate x2 into the model where x0 and x3 were already present, SSreg(β1|β0,β3,β2) defines the extra variability explained due to the addition of the covariate x1 into the model where x0, x3, and x2 were already present.

  • We create some models and then apply the anova() routine. model1 = lm(y~x0+x3) model2 = lm(y~x0+x3+x2) anova(model1, model2) # This will give the extra sum of squares for adding x2 into the model with x0 & x3 model3 = lm(y~x0+x3+x2+x1) anova(model2, model3) #This will give the extra sum of squares for adding x1 into the model with x0, x3, x2

Related Solutions

We consider the multiple linear regression with LIFE (y) as the response variable, and MALE, BIRTH,...
We consider the multiple linear regression with LIFE (y) as the response variable, and MALE, BIRTH, DIVO , BEDS, EDUC, and INCO, as predictors. QUESTION: Plot the standardized residuals against the fitted values. Are there any notable points. In particular look for points with large residuals or that may be influential. # please screenshot the Rcode for the plot. # data information are as follows: "STATE" "MALE" "BIRTH" "DIVO" "BEDS" "EDUC" "INCO" "LIFE" AK 119.1 24.8 5.6 603.3 14.1 4638...
Ten observations on the response variable y associated with two regressor variables x1 and x2 are...
Ten observations on the response variable y associated with two regressor variables x1 and x2 are given in the following table. The model fitted to these observations is yi = β 0 + β 1x1i + β 2x2i + εi , i = 1, …, n , (3) where ε’s are identically and independently distributed as a normal random variable with mean zero and a known variance σ ^2 = 3. Observation # y x1 x2 1 7 9 1...
Consider a binary response variable y and two explanatory variables x1 and x2. The following table...
Consider a binary response variable y and two explanatory variables x1 and x2. The following table contains the parameter estimates of the linear probability model (LPM) and the logit model, with the associated p-values shown in parentheses. Variable LPM Logit Constant −0.40 −2.20 (0.03 ) (0.01 ) x1 0.32 0.98 (0.04 ) (0.06 ) x2 −0.04 −0.20 (0.01 ) (0.01 ) a. At the 5% significance level, comment on the significance of the variables for both models. Variable LPM Logit...
For the table below, if Y is the dependent variable and X1 and X2 are the...
For the table below, if Y is the dependent variable and X1 and X2 are the independent variables. Using the linear regression equation Y=-0.45X1-1.34X2+15.67, which observation has the largest absolute residual? Observation number Actual Y x1 X2 1 4.5 6.8 6.1 2 3.7 8.5 5.1 3 5 9 5 4 5.1 6.9 5.4 5 7 8 4 6 5.7 8.4 5.4 The first observation The third observation The fifth observation The second observation
For the table below, if Y is the dependent variable and X1 and X2 are the...
For the table below, if Y is the dependent variable and X1 and X2 are the independent variables. Using the linear regression equation Y=-0.45X1-1.34X2+15.67, find the Sum of Squared Residuals? (choose the best answer) Observation number Actual Y x1 X2 1 4.5 6.8 6.1 2 3.7 8.5 5.1 3 5 9 5 4 5.1 6.9 5.4 5 7 8 4 6 5.7 8.4 5.4 2.57 2.97 3.2 3.5
Air Transporation model Model Y = bo + b1* x1 + b2 * x2 For the...
Air Transporation model Model Y = bo + b1* x1 + b2 * x2 For the 1st case, the Dependent variable is:- Revenue passenger per mile This variable is dependent on the independent variables like [ load factor and enplanements For 2nd case, the Dependent variable is:- Load factor This variable is dependent on the two independent variables like [ available seats and enplanement ] What is the model assumptions? Also how these models related to past economic theory on...
A regression model of the form y = beta0 + beta1 x1 + beta2 x2 +...
A regression model of the form y = beta0 + beta1 x1 + beta2 x2 + beta3 x3 + E was built using 20 observations. Partially completed regression output tables are provided below. What are the values of A, B, and C? Table 1 Statistic Value R-Square A Adjusted R-Square B Standard Error (RMSE) C n 20 Table 2 Source DF SS MS F P-Value Regression D 175 H J K Error E G I Total F 250 A regression...
           Case             Y           X1           X2
           Case             Y           X1           X2           X3           X4           X5           X6 1 43 45 92 61 39 30 51 2 63 47 73 63 54 51 64 3 71 48 86 76 69 68 70 4 61 35 84 54 47 45 63 5 81 47 83 71 66 56 78 6 43 34 49 54 44 49 55 7 58 35 68 66 56 42 67 8 74 41 66 70 53 50...
Prove E(X1 + X2 | Y=y) = E(X1 | Y=y) + E(X2 |Y=y). Prove both cases...
Prove E(X1 + X2 | Y=y) = E(X1 | Y=y) + E(X2 |Y=y). Prove both cases where all random variables are discrete and also when all random variables are continuous.
Using Y as the dependent variable and X1, X2, X3, X4 and X5 as the explanatory...
Using Y as the dependent variable and X1, X2, X3, X4 and X5 as the explanatory variables, formulate an econometric model for data that is (i) time series data (ii) cross-sectional data and (iii) panel data – (Hint: please specify the specific model here not its general form).
ADVERTISEMENT
ADVERTISEMENT
ADVERTISEMENT