Extensions of the linear regression

The standard linear regression model provides interpretable results and works quite well on many real-world problems. However, it makes several highly restrictive assumptions that are often violated in practice.Two of the most import assumptions state that the relationship between the predictors and response are additive and linear.

  • additive assumption: the effect of changes in a predictor on the response is independent of the values of the other predictors.
  • linear assumption: the linear assumption states that the change in the response due to a one-unit change in is constant, regardless of the value of

Removing the Additive Assumption

The linear models that formed the basis for this conclusion assumed that the effect on sales of increasing one advertising medium is independent of the amount spent on the other media. However, this simple model may be incorrect.Suppose that spending money on radio advertising actually increases the effectiveness of TV advertising, so that the slope term for TV should increase as radio increases. One way of extending this model to allow for interaction effects is to include a third predictor, called an interaction term, which is constructed by computing the product of and . This results in the model: It is sometimes the case that an interaction term has a very small p-values, but the associated main effects(like TV and radio) do not.

The hierarchical principle states that if we include an interaction in a model, we should also include the main effects, even if the p-values associated with their coefficients are not significant.

In other words, if the interaction between and seems important, then we should include both and in the model even if their coefficient estimates have large p-values.

The concept of interactions applies to quantitative variables, as well to qualitative variables, or to a combination of quantitative and qualitative variables. In fact, an interaction between a qualitative variable and a quantitative variable has a particularly nice interpretation.

Non-linear Relationship

The linear regression model assumes a linear Relationship between the response and predictors. But in some cases, the true Relationship between the response and predictors maybe non-linear. We present a very simple way to directly extend the linear model to accommodate non-linear relationships, using polynomial regression. The following formula is a model: The formula is using a non-linear function of horsepower, but it's still a linear model! That is, it's simply a multiple linear regression with and . So we can use standard linear regression software to estimate in order to produce a non-linear fit.

Potential Problems

Some problems may occur. Most common among these are the following:

  1. non-linearity of the response-predictor relationships
  2. correlation of error terms
  3. non-constant variance of error terms
  4. outliers
  5. high-leverage points
  6. collinearity

non-linearity of the Data

Residual plots are a useful graphical tool for identifying non-linearity. If the residual plot indicates that there are non-linear associations in the data, then a simple approach is to use non-linear transformations of the predictors, such as , in the regression model.

correlation of error terms

Such correlations frequently occur in the context of time series data, which consists of observations for which measurements are obtained at discrete points in time.In many cases, observations that are obtained at adjacent time points will have positively correlated errors. In general, the assumption of uncorrelated errors is extremely import for linear regression as well as for other statistical methods, and good experimental design is crucial in order to mitigate the risk of such correlations.

non-constant variance of error terms

It is often the case that the variances of the error terms are non-constant. When faced with this problems, one possible solution is to transform the response Y using a concave function such as . Such a transformation results in a greater amount of shrinkage of the larger response, leading to a reduction in heteroscedasticity.

outliers

An outliers is a point for which is far from the value predicted by the model. It's typical for an outliers that does not have an unusual predictor value to have little effect on the least squares fit. However, maybe RSE will decrease, since the RSE is used to compute all confidence intervals and p-values, such a dramatic increase caused by a single data point can have implications for the interpretation of the fix, similarly, inclusion of the outlier causes the to decline.

How to identify outliers? To address this problem, insteam of plotting the residuals, we can plot the studentized residuals, computed by dividing each residual by its estimated standard error.observations whos studentized residuals are greater than 3 in absolute value are possible outliners.

care should be taken to remove outliers, since an outlier may instead indicate a deficiency with the model, such as a missing predictor.

We just saw that outliers are observations for which the response is unusual give the predictor , in contrast, observations with high leverage have an unusual value for . Removing the high leverage observation has a much more substantial impact on the least squares line than removing the outliner. In order to quantify an observation's leverage, we compute the leverage statistic. A large value of this statistic indicates an observation with high leverage: The leverage statistic is always between 1/n and 1, and the average leverage for all the observations is always equal to .So if a given observation has a leverage statistic that greatly exceeds , then we may suspect that the corresponding point has high leverage.

A simple way to detect collinearity is to look at the correlation matrix of the predictors. An element of this matrix that is large in absolute value indicates a pair of highly correlated variables, and therefore a collinearity problem in the data.

But for multicollinearity(collinearity to exist between three or more variables even if no pair of the variables has a particularly high correlation), a better way to access multicollinearity is to compute the variance inflation factor(VIF).

When faced with the problem of collinearity, there are two simple solutions:

  1. drop one of the problematic variables from the regression.
  2. combine the collinear variables together into a single predictor.

results matching ""

    No results matching ""