Logistic Regression
- A supervised learning algorithm used for classification problems
- Sigmoid Function: values range from 0 - 1
- Assumptions
- Response Variable is Binary
- Independence - observations in the dataset are independent of each other
- No Multicollinearity - none of the predictor variables are highly correlated with each other
- No Extreme Outliers
- Linear Relationship between Explanatory Variables and Logit of Response Variable
- Sample Size is Sufficiently Large
Pros | Cons |
---|---|
Highly interpretable | Can overfit with small, high-dimensional data |
Applicable for multi-class predictions |
- Multiclass Classification: One-vs-Rest (OvR)
- Approaches
- Gradient Descent
- Binary Cross Entropy / Log Loss
- Cost(hθ(x),y) = −y log(hθ(x))−(1−y) log(1−hθ(x))
- Gradient Descent
- Model Performance Evaluation
- Confusion Matrix, Accuracy, Precision, Recall, F1-Score, ROC Curve