Exam 5: Multiple Regression

arrow
  • Select Tags
search iconSearch Question
  • Select Tags

Weighted least-squares is efficient when E[ui2] \neq E[uj2] = σ\sigma 2 for all i \neq j.

Free
(True/False)
4.9/5
(33)
Correct Answer:
Verified

True

AVG =77 RANGE: 30-100 A - 4 B - C - 2 D - F - 2 -Draw a diagram indicative of negative serial correlation. Be sure to label the axes.

Free
(Essay)
4.7/5
(37)
Correct Answer:
Verified

AVG =77 RANGE: 30-100 A - 4 B - C - 2 D - F - 2 -Describe the Hildreth-Lu Scanning procedure.

Free
(Essay)
4.8/5
(41)
Correct Answer:
Verified

Since ρ\rho is expected to lie between plus and minus unity, a computer can easily check to see which value of ρ\rho gives the best results in a GLS regression.Assume ρ\rho = -1; Run GLS; Assume ρ\rho = -.9; Run GLS; Assume ρ\rho = -.8; Run GLS; ...; Assume ρ\rho = +.9; Run GLS; Assume ρ\rho = +1; Run GLS. Of all these GLS regressions, use the one that has the best fit according to the R2 or SER.

AVG =77 RANGE: 30-100 A - 4 B - C - 2 D - F - 2 -Describe the weighted least-squares procedure.

(Essay)
4.9/5
(38)

AVG =77 RANGE: 30-100 A - 4 B - C - 2 D - F - 2 -Describe the maximum likelihood procedure.

(Essay)
4.7/5
(38)

The Durbin-Watson statistic is invalid in autoregressive models, models without a constant term, and models with n = 9.

(True/False)
4.8/5
(43)

R-squared from a GLS regression is directly comparable to the R-squared from the same regression estimated using OLS.

(True/False)
4.8/5
(47)

Serial correlation is when E[ui uj] = 0 for all i \neq j.

(True/False)
4.9/5
(37)

R-squared is biased downward in a regression suffering from serial correlation.

(True/False)
4.7/5
(38)

AVG =77 RANGE: 30-100 A - 4 B - C - 2 D - F - 2 -Given Yi = β^0\hat { \beta } _ { 0 } + β^1\hat { \beta } _ { 1 } X1i + β^2\hat { \beta } _ { 2 } X2i + ei, describe the White test for heteroskedasticity making sure to specify the auxiliary regression in this case.

(Essay)
4.7/5
(43)

A regression with a Durbin-Watson statistic close to 4 most likely suffers from negative autocorrelation.

(True/False)
4.8/5
(43)

Application of the Newey-West technique will alter the estimates of the P-values on the structural parameters.

(True/False)
4.7/5
(28)

AVG =77 RANGE: 30-100 A - 4 B - C - 2 D - F - 2 -What do you recommend in a situation where GLS does not remedy autocorrelation? Explain.

(Essay)
4.8/5
(40)

The error terms (ut) from a regression are white noise when ut ~ N(0, σ\sigma 2).

(True/False)
4.7/5
(35)

AVG =77 RANGE: 30-100 A - 4 B - C - 2 D - F - 2 -Are the standard errors of the coefficients in a regression expected to increase or decrease after Newey-West is applied? Explain why?

(Essay)
4.8/5
(41)

In the presence of autocorrelated error terms, GLS yields BLUE parameter estimates.

(True/False)
4.9/5
(40)

AVG =77 RANGE: 30-100 A - 4 B - C - 2 D - F - 2 -What is generalized least-squares?

(Essay)
4.8/5
(43)

AVG =77 RANGE: 30-100 A - 4 B - C - 2 D - F - 2 -Describe 3 ways to estimate ? in a generalized least-squares regression.

(Essay)
4.9/5
(35)

Autocorrelation increases the probability of a TYPE II error on a test of significance on a given structural parameter.

(True/False)
4.8/5
(34)

AVG =77 RANGE: 30-100 A - 4 B - C - 2 D - F - 2 -Given Yi = β^0\hat { \beta } _ { 0 } + β^1\hat { \beta } _ { 1 } X1i + β^2\hat { \beta } _ { 2 } X2i + β^2\hat { \beta } _ { 2 } X3i + ei, write the auxiliary regression required to perform the Park test for heteroskedasticity.

(Essay)
4.7/5
(35)
close modal

Filters

  • Essay(0)
  • Multiple Choice(0)
  • Short Answer(0)
  • True False(0)
  • Matching(0)