A student stated: “Adding predictor variables to a regression model can never reduce \(R^2\), so we should include all available predictor variables in the model.” Comment.
Why is it not meaningful to attach a sign to the coefficient of multiple correlation \(R\), although we do so for the coefficient of simple correlation \(r_{12}\)?
In a small-scale regression study, the following data were obtained,
Y | \(X_1\) | \(X_2\) |
---|---|---|
42.0 | 7.0 | 33.0 |
33.0 | 4.0 | 41.0 |
75.0 | 16.0 | 7.0 |
28.0 | 3.0 | 49.0 |
91.0 | 21.0 | 5.0 |
55.0 | 8.0 | 31.0 |
Assume the standard multiple regression model with independent normal error terms. Compute \(\mathbf{b}\), \(\mathbf{e}\), \(\mathbf{H}\), SSErr, \(R^2\), \(s^2_{\mathbf{b}}\), \(\hat Y\) for \(X_1 = 10\), \(X_2 = 30\). You can do the computations using software or by hand, although it would be lengthy to do them by hand.
The data set “Brand preference” is available here.
It was collected to study thee relation between degree of brand liking (Y, first column) and moisture content (\(X_1\), second column) and sweetness (\(X_2\), third column) of the product.