A quick review on logsitc regression
Logistic regression tries to model the relationship between predictors and the conditional disrtibution of the response Y given the predictors X using logistic function.
logistic regression also has several assumptions
- The predictors are linearly associated with log odds
- The response is binary variable with only two classes
To estimate the unknown parameters of the function, ,maximum likelihood is used instead of the least squares. We aim to fit a line that has the maximum likelihood after converting this straight line into a s-shaped curve using logistic function of log odds.
Sometimes, we may need to classify the response with more than two classes; in other words, we need to perform multiclass classification. Although the two-class logistic regression model can be extended to multiple-class extentions, but in practice they tend not to be used very often because another method, Linear Discriminant Analysis is more popular for multiple-class classification.
Linear Discriminant Analysis and Quadratic Discriminant Analysis
Agenda
- LDA
- QDA
- How to choose between LDA and QDA (variance/bias trade off)
LDA
What is Linear Discriminant Analysis?
Unlike logistic regression, in which we directly model the conditional distribution of response given predictors X, LDA indirectly models the conditional distribution of the response by using Bayes' theorem. In other words, the method models separately the distribution of predictors X given Y and then uses Bayes theorem to flip these around into estimate for . If the observations are assumed to be normal, then the model is similar to logistic regression model.
How Bayes Theorem works for LDA?
Before explaining the use of Bayes Theorem in LDA, let me introduce some terminology about Bayes Theorem first:
Prior Probability
Prior probability, in Bayes statistical inference, is the probability of an event before new information or data is collected. It is the best rational assessment of probability of an outcome based on current knowledge before an experiment is performed.
Posterior Probability
Prior probability needs to be revised when new information or data becomes available. The revised probability is Posterior probability. In statistical terms, it is the probability of event A occurring given event B occurring.
Bayes Theorem
Bayes theorem describes the probability of an event based on new information that is related to the event, that is posterior probability.
Now suppose that we want to classify the observation into one of k classes, where . We let
represent the prior probability and
denote the probability distribution that an observation comes from Kth class. We have:
-
is the prior probability
-
is the posterior probability
-
is Bayes Theorem
Estimating parameters of LDA
First let's look at the function of LDA
We do not need to understand and memorize the complex formula. All we need to know is that this is a linear function of x and it has two unknown parameters, and
. This is why we call the statistical method Linear.
Based on the formula, we can get the assumption of Linear Discriminant Analysis: LDA assumes that observations come from Gaussian distribution with a class-specific mean vector and a common variance
The function is a little bit complicated because of Greek Letters. I think it would be better if it is written like this:
In fact, I have written that LDA is just another indirect method of modeling the distribution of response Y compared with logistic regression. So it is very similar to logistic regression and they just differ in the fitting procedure. Therefore, LDA is still modeling the relationship between log odds and predictors, but LDA needs to estimate and
from sample while logistic regression needs to estimate
using maximum likelihood.
Therefore, the equation above can be understanded as
-
corresponds to
-
corresponds to
The estimation of mean and variance of sample is pretty easy, we just need to calculate them based on the training observations. And we want the response to be the largest when plugging the estimates for mean and variance into the equation.
OK, now we know a lot about LDA, let me compare it with logistic regression.
Logistic regression Versus LDA
Consider the two-class setting with p=1 estimator, and let and
be the probabilities that x belongs to class 1 and class2.
- Logistic regression can be wirtten as
- For LDA, the model is
Therefore, based on the functional form of f, we can find that both LDA and logistic regression produce linear decision boundaries.
On the other hand, there is also a significant difference between the two approaches. They have different method of estimating parameters.
Logistic regression estimate parameters using maximum likelihood, while linear discriminant analysis estimate parameters, and
, using the estimated mean and variance from a normal distribution.
When to use LDA and when to use logistic regression?
On the condition that the decision boundary is linear, we should choose between the two approaches based on:
- If the observations come from Gaussian distribution with specific class mean and and a common covariance matrix in each class, LDA can provide improvements over logistic regression
- If the normal assumption is not met, then logistic regression performs better than LDA
- If the sample is very small and the response has more than two classes, we also should choose LDA
Quadratic Discriminant Analysis
LDA assumes that observations within each class are drawn from a multivariate Gaussian distribution with a class-specific mean vector and covariance matrix that is common to all K classes. In contrast, Quadratic discriminant analysis assumes that each class has its own covariance matrix: perhaps the correlation between predictors and class 1 is 0.5 while the correlation between predictors and class 2 is -0.5. And the fuctional form of QDA is quadratic instead of linear, and this is how quadratic discriminant analysis is called quadratic.
The function of QDA
Again we need to understand not the complex equation but that it is a quadratic function of x and it also has two unknown parameters
Summary
So based on this, we can summarize the similarities and differences between LDA and QDA.
- LDA is a linear function with linear decision boundary, while QDA is a quadratic function with quadratic decision boundary
- LDA assumes that each class has the same covariance matrix while QDA assumes that each class has its own covariance matrix
- LDA and QDA both assume that observations come from Gaussian distribution
What are the scenarios where we should LDA and the ones for QDA?
The question depends on one core conception that is omnipresent in the studying and career of machine learning, variance and bias tradeoff.
- If K classes share the common covariance matrix, the LDA has a linear decision boundary, which means that the coefficients of LDA model should be linear. In this setting, LDA is a less flexible classifier than QDA and thus has lower variance. So LDA can improve the performance of predictions. In addition, if the training observations are very small, LDA is also better option because reducing variance is the priority. On the other hand, QDA will only fits a model much more flexible than necessary, meaning substantially high variance that even cannot be offset by the decrease in bias.
- If the training set is very large or if the assumption of same covariance matrix is not held, QDA is recommended. In the case of large train set, the variance reduces as more samples come in and so is not a problem for model. So we need to pay attention to bias and a more flexible model is needed.
Given that variance/bias tradeoff lies at the heart of machine learning and statistical learning, Let me introduce it. As long as we grasp the essence of it, we basically know how to choose among various machine learning algorithms.
Variance/Bias tradeoff
We know that the performance of model can be evaluated by MSE(mean squared error), given by:
We want the MSE to be as small as possible because small mse means that the predicted value is close to true value.
Also, we know that the evaluation of model should be performed on test set rather than train set. That is to say we finally want to have a model with the minimum test mse.
Now the test MSE can be decomposed into the sum of three parts:
The equation tells us that in order to minimize the expected test error, we need to select a statistical learning method that simultaneously achieves low variance and low bias.
Let me quote one example figure from Introduction to Stats Learning for explanation.
First, what is variance?
Variance refers to the amount by which f would change if we estiamted it using different training data set. Since the training data are used to fit the statistical learning method, different training data sets will result in a differnt f. But ideally the estimator for f should not vary too much between training sets. However, if a method has high variance, then a small change in training data can cause large changes in f. Takes the green curve from the left-hand panel of the figure as an example, the green curve is following the data very closely, which means that the predicted values are very close to the actual value. So this curve has low bias but high variance.
Why does it have high variance? The reason is that changing any data points may cause the estimate of f change considerably. We can conclude this by observing the slope of the green curve.
What is bias?
Bias refers to the error that is introduced by approximating a real-life problem by a much simpler model. For example, linear regression assumes that there is a linear relationship between X and Y. It is unlikely that any real-life problem truly has such a simple relationship, and so performing linear regression will undoubtlessly lead to inaccurate predicted values that are far away from true values, that is some bias in f.
Now with the knowledge of variance and bias, What is variance/bias tradeoff?
Good test set performance of a statistical learning method requires low variance and low bias. This is referred to as a trade-off because it is easy to obtain a method with extremely low bias and high variance (for example we fit a curve that passes through every single observation) or high bias and low variance (we just fit a horizontal line).
A general rule about v/b tradeoff
the variance increases but bias decreases as flexibility increases. And the relative rate of change of the two terms determine whether the overall test mse increase ot not. As we increase the flexibility of statistical methods, the bias tends to reduce faster than the variance increases, so the overall test MSE decreases. However, at some point increasing the flexibility has little impact on the reduction in bias but significantly increases variance of model. In this case, the overall test MSE increases. This is why sometimes a simple linear regression model performs better than highly flexible methods, such as KNN and decision trees.
Exercise
OK, we know the conception of variance/bias tradeoff, now let me post one exercise about the conception of variance/bias tradeoff.
- Provide a sketch of typical (squared) bias, variance, training error, test error, and Bayes (or irreducible) error curves, on a single plot, as we go from less flexible statistical learning methods towards more flexible approaches. The x-axis should represent the amount of flexibility in the method, and the y-axis should represent the values for each curve. There should be five curves. Make sure to label each one.
- Explain why?
Solution
- Trian Error. Train error is a monotonically decreasing blue curve because the curve will more closely fit data points as flexibility increases, meaning more accurate estimate of f.
- Test Error. Test error is an U-shaped curve, monotonically decreasing but starting to significantly increase after a pariticular point. The reason for this is that test error decreases as flexibility increases but at some point increasing flexibility has little impact on bias but significantly increases the variance. In this setting, we may have problem of overfitting, training mse being small but test mse being large.
- Bias and Variance. As a general rule, as flexibility increases, bias will decrease and variance will increase. Variance refers to the amount by which f would change if we estimated it using different training set. Bias refers to the error that is introduced by approximating a real-life problem by using a simpler model. As flexibility increases, the highly flexible learning methods will follow the data very closely, and in this case they have low bias but high variance because a small change in data points can cause substantial changes in estimates of f.
- Bayes error. This is a horinzontal line intersecting at the y axis and below Test MSe because the expected test MSE will always be greater the Var(ε*)
Next article
- The comparison of various classifiers
- Confusion matrix for evaluation of the performance of classifiers