cs-5821/hw2/answers

182 lines
8.4 KiB
Plaintext
Raw Permalink Normal View History

2017-01-30 06:45:45 +00:00
2017-02-02 23:54:50 +00:00
1. Describe the null hypotheses to which the p-values given in Table
3.4 correspond. Explain what conclusions you can draw based on
these p-values. Your explanation should be phrased in terms of
sales , TV , radio , and newspaper , rather than in terms of the
coefficients of the linear model.
P-values that are very small indicate that the model for that
predictor is likely to account for a significant amount of the
association between the predictor and the response. If that is
true, then, we reject the null hypothesis, and conclude that a
relationship exists between the predictor and the response. The
p-values computed from the response of sales to marketing budget
for each marketing paradigm indicate will give us insight into
which of these predictors have a strong relationship with sales
of this product.
2017-02-03 04:47:24 +00:00
TV marketing and radio marketing both have a strong relationship
to sales, according to their linear regression p-values, but
newspaper advertising does not appear to be effective, given
that the linear model does not account for much of the variation
in sales across that domain. We can conclude that cutting back
on newspaper advertising will likely have little effect on the
sales of the product, and that increasing TV and radio
advertising budgets likely will have an effect. Furthermore, we
can see that radio advertising spending has a stronger
relationship with sales, as the best-fit slope is significantly
more positive than the best fit for TV advertising spending, so
increasing the radio advertising budget will likely be more
effective.
2017-02-02 23:54:50 +00:00
3. Suppose we have a data set with five predictors, X₁ = GPA, X₂ =
IQ, X₃ = Gender (1 for Female and 0 for Male), X₄ = Interaction
between GPA and IQ, and X₅ = Interaction between GPA and Gender.
The response is starting salary after graduation (in thousands of
dollars). Suppose we use least squares to fit the model, and get
β₀ = 50, β₁ = 20, β₂ = 0.07, β₃ = 35, β₄ = 0.01, β₅ = 10.
2017-01-30 06:45:45 +00:00
2017-02-03 04:47:24 +00:00
This is the model: ŷ = 50 + 20 X₁ + 0.07 X₂ + 35 X₃ + 0.01 X₄ +
-10 X₅
2017-01-30 06:45:45 +00:00
2017-02-03 04:47:24 +00:00
For fixed IQ and GPA, we can infer that the starting salary
for a female sharing an IQ and GPA with her male counterpart
will make (35*1 - 10*(GPA*1)) more starting salary units
than her male counterpart. This means that at very low
GPAs(maybe this includes people who didn't attend school?),
males have a lower starting wage, and as GPA grows, males
make a larger starting salary from that point, overtaking
females at GPA=3.5. Therefore,
2017-01-30 06:45:45 +00:00
2017-02-03 04:47:24 +00:00
(a) Which answer is correct, and why? → iii. For a fixed value
of IQ and GPA, males earn more on average than females
provided that the GPA is high enough.
This one is correct.
2017-01-30 06:45:45 +00:00
2017-02-03 04:47:24 +00:00
(b) Predict the salary of a female with IQ of 110 and a GPA of
4.0.
2017-01-30 06:45:45 +00:00
2017-02-03 04:47:24 +00:00
ŷ = 50 + 20*4.0 + 0.07*110 + 35*1 + 0.01*(4.0*110) - 10*(4.0*1)
2017-01-30 06:45:45 +00:00
2017-02-03 04:47:24 +00:00
→ ŷ = 137.1 salary units
2017-01-30 06:45:45 +00:00
2017-02-03 04:47:24 +00:00
(c) True or false: Since the coefficient for the GPA/IQ
interaction term is very small, there is very little evidence of
an interaction effect. Justify your answer.
2017-01-30 06:45:45 +00:00
2017-02-03 04:47:24 +00:00
False. There is still a noticeable effect because the
coefficient for IQ's effect alone is only 7 times greater
than the coefficient of the interaction term. So, this term
holds significant weight compared to the overall
response of the model to IQ.
2017-01-30 06:45:45 +00:00
2017-02-02 23:54:50 +00:00
4. I collect a set of data (n = 100 observations) containing a
single predictor and a quantitative response. I then fit a linear
regression model to the data, as well as a separate cubic
2017-02-03 04:47:24 +00:00
regression, i.e. Y = β₀ + β₁ X + β₂ X² + β₃ X³ + ε.
2017-01-30 06:45:45 +00:00
2017-02-02 23:54:50 +00:00
(a) Suppose that the true relationship between X and Y is
2017-02-03 04:47:24 +00:00
linear, i.e. Y = β₀ + β₁ X + ε. Consider the training residual
2017-02-02 23:54:50 +00:00
sum of squares (RSS) for the linear regression, and also the
training RSS for the cubic regression. Would we expect one to be
lower than the other, would we expect them to be the same, or is
there not enough information to tell? Justify your answer.
2017-01-30 06:45:45 +00:00
2017-02-03 04:47:24 +00:00
For the training data, the cubic regression might return a
better RSS than the linear regression, but this would only
be because the cubic is fitting points that are varied
according to the ε random error. It also may not, depending
on how that random error expressed itself in this case.
2017-01-30 06:45:45 +00:00
(b) Answer (a) using test rather than training RSS.
2017-02-03 04:47:24 +00:00
For the test error, the RSS will almost certainly be greater
for the cubic model than the linear model, because the
random error ε will likely express itself in a way that is
inconsistent with the noise that the cubic model adopted
during its training. The linear model will be more likely to
have a lower RSS the more test data is used against the
models.
2017-02-02 23:54:50 +00:00
(c) Suppose that the true relationship between X and Y is not
linear, but we dont know how far it is from linear. Consider
the training RSS for the linear regression, and also the
training RSS for the cubic regression. Would we expect one to be
lower than the other, would we expect them to be the same, or is
there not enough information to tell? Justify your answer. (d)
2017-02-03 04:47:24 +00:00
Answer (c) using test rather than training RSS.
The cubic model will pick up more information because of its
additional degrees of freedom. If the true relationship is
more complex than linear, then the cubic model will likely
have a lower RSS over the linear model. If the model is less
complex than linear (E.G. perhaps it is just a constant
scalar relationship) then the linear model will still be
more likely to have a smaller RSS, because the cubic will
again pick up information from the ε noise that is not
inherent in the real relationship.
8. This question involves the use of simple linear regression on the Auto
data set.
(a) Use the lm() function to perform a simple linear regression
with mpg as the response and horsepower as the predictor. Use
the summary() function to print the results. Comment on the
output. For example:
There is definitely a correlation between horsepower and
mpg. The RSE is ~4.9, which is not insignificant and does
indicate that the response may not be truly linear, but it
is small enough relative to the mpg magnitude that it's
clear a relationship exists. The R² statistics corroborates
this by indicating (it has a small value at ~0.6) that a
large proportion of the mpg variability is explained by the
model. mpg has a negative correlation with horsepower,
indicated by the negative coefficient on the horsepower
factor.
For example, for a vehicle with 98 horsepower, one can
expect with 95% confidence that the mpg will be within 23.97
and 24.96, if the vehicles follow our model. However, after
incorporating the irreducible error, the prediction turns
out to be much less precise, with a 95% prediction interval
spanning 14.8 to 34.1. Some of this variability may also be
reduced by using a quadratic model, from visual inspection
of the plot.
(b) Plot the response and the predictor. Use the abline() function
to display the least squares regression line.
Attached.
(c) Use the plot() function to produce diagnostic plots of the least
squares regression fit. Comment on any problems you see with
the fit.
Attached. From these four plots it's clear there is a lot of
variability that remains unexplained by the linear model.
The standardized residuals plotted against the fitted values
shows clearly that the variability is strong, with values
consistenly lying outside 1 standardized residual unit, but
still within a tight range that doesn't extend past 3, which
is often considered an approximate threshold to indicate
values that aren't explained well by the model. There are
many points with high leverage, and these values have less
residual by default, of course, and in both of these graphs
we see a few points (323, 330) that are rearing their ugly
heads. These seems to be the bit of "uptick" toward the
higher end of the horsepower scale that would probably be
picked up by a quadratic fit.