The logistic model

How should we model the relationship between and ? In logistic regression, we use the logistic function:

  • for low balances, we now predict the probability of default as close to, but never close to, zero.
  • for high balances, we predict a default probability close to ,but never above, one.

The quantity is called the odds, and can take on any value between 0 and . Odds are traditionally used instead of probabilities in horse-racing, since they relate more naturally to the correct betting strategy.

estimating the regression coefficients

The coefficients are unknown, and must be estimated based on the available training data. In Chapter 3, we used the least squares approach to estimate the unknown linear regression coefficients. The more general method of maximum likelihood is preferred, since it has better statistical properties.

The basic intuition behind using maximum likelihood to fit a logistic regression model is as follows: we seek estimates for and such that the predicted probability of default for each individual, using previous formula, corresponds as closely as possible to the individual's observed default status.

This intuition can be formalized using a mathematical equation called a likelihood function: The estimates and are chosen to maximize this likelihood function.

Maximum likelihood is a very general approach that is used to fit many of the non-linear models that we examine throughout this book. In the linear regression setting, the least squares approach is in fact a special case of maximum likelihood.

Making predictions

Once the coefficients have been estimated, it is a simple matter to compute the probability of default for any given credit card balance.

One can use qualitative predictors with the logistic regression model using the dummy variable approach from Section 3.3.1.

Multiple logistic regression

We use the maximum likelihood method to estimate .

A student is riskier than a non-student if no information about the student's credit card balance is available. However, that student is less risky than a non-student with the same credit card balance!

This simple example illustrates the dangers and subtleties associated with performing regressions involving only a single predictor when other predictors may also be relevant.

Logistic Regression for > 2 Response Classes

Discriminant analysis, is popular for multiple-class classification.

References

最大似然估计和最小二乘法

从odds理解LR

results matching ""

    No results matching ""