-
Couldn't load subscription status.
- Fork 0
Logistic Regression
Niranjan edited this page Aug 31, 2017
·
4 revisions
- Used when dependent variable (response) is categorical (usually 2 classes). Extensions include
multinomial logistic regressionfor > 2 classes andordered logistic regressionfor ordered responses,mixed logit,conditional random fields,conditional logistic regression, etc. - Used to estimate probability of binary response using 1 or more covariates
-
P(Y|X)is Bernoulli, not Gaussian. -
Y_iare not identically distributed sinceP(Y_i|X_i)depends on the value ofX_i. ButY_iare independent conditional onX_iand$\beta$ - Predicted values are probabilities (in
[0,1]due to the logistic function); threshold predicted probabilities to classify predictions into categories - Alternative to
linear discriminant analysis. -
Logitfunction is inverse ofLogisticfunction.Logit/log odds of probability is equal to RHS of linear regression equation.
Model parameters must be estimated via iterative method, no closed form expression available like in linear regression. Failure of method to converge can occurs due to:
-
p >> n=> conservative Wald statistic => non-convergence - Mulicollinearity => high std errors of model parameters
- Sparseness in data => large number of empty cells => problematic for categorical data => no convergence because log(0) is undefined => collapse categories or add constant to all cells
- Complete separation => all predictions are accurate => errors present
- Use
devianceto assess goodness of fit of model; analogous toR^2in linear regression. Small values => less deviance from 'full' model - Pseudo R^2 - several measures available, each with its own limitations
- Likelihood Ratio test
- Wald statistic
ISLR, Section 4.3- Wikipedia
-
Linear Methods
- Linear Regression
- Logistic Regression
- Stochastic Gradient Descent
-
Shrinkage methods
- Ridge Regression
- LASSO
- LARS
-
Derived Input Methods
- Principal Components Regression
- Partial Least Squares Regression
-
Non-linear methods
- MARS
- Polynomial Regression
- LOESS
- Splines
- Generalized Additive Models (GAMS)
- Isotonic Regression
- Bayes error rate
- k-Nearest-Neighbors
- Linear Discriminant Analysis
- QDA
- Support Vector Machines
Tree-based Methods
Model Selection / Assessment / Resampling
- Bias-Variance trade-off
- BIC
- MDL
- Vapnik-Chervonenkis Dimension
- Cross-validation
- Bootstrap
Clustering
- K-Means
- Gaussian Mixture Models
- Hierarchical Clustering
- Affinity Propagation
- Mean Shift
- Spectral Clustering
- DBSCAN
- Birch
- Power Iteration Clustering
Dimensionality Reduction
- PCA
- SVD
- ICA
- Factor Analysis
Density Estimation
Frequent Pattern Mining
Recommender Systems
NLP