Charmi Chokshi
April 08, 2020
68

# Logistic Regression

April 08, 2020

## Transcript

2. ### Let’s Start Basics of Machine Learning! I’m, Charmi Chokshi An

ML Engineer at Shipmnts.com and a passionate Tech-speaker. A Critical Thinker and your mentor of the day! Let’s connect: @CharmiChokshi

7. ### Regression and Classification A regression model predicts continuous values. For

example, regression models make predictions that answer questions like the following: • What is the value of a house in California? • What is the probability that a user will click on this ad? A classification model predicts discrete values. For example, classification models make predictions that answer questions like the following: • Is a given email message spam or not spam? • Is this an image of a dog, a cat, or a hamster?

9. ### Remember? By the end of the session, you will have

run a machine learning experiment to classify foods as pizza or not pizza Not-Pizza Pizza
10. ### The Acceptance Dilemma My Story: I am highly interested in

pursuing M.S. in Artificial Intelligence at UGoog. My GRE Score is 315 and I fancy my chances of getting in. As a prospective graduate student, I start browsing through the GRE scores of students that had been admitted to UGoog and those which were rejected. I could predict my chances through regression model. Unless the graph looks…...

losses

13. ### The Solution: Logistic Regression Many problems require a probability estimate

as output. Logistic regression is an extremely efficient mechanism for calculating probabilities. You might be wondering how a logistic regression model can ensure output that always falls between 0 and 1. As it happens, a sigmoid function, defined as follows, produces output having those same characteristics:
14. ### Log Loss Now, z = w 0 + x 1

w 1 , and we take a log loss function- The weights are then updated by the log loss until the the log loss converges to almost 0. My regression model now gives a probability of my acceptance at UGoog as an output.
15. ### Regression, Thresholding and Classification Logistic regression returns a probability. You

can use the returned probability as it is or convert the returned probability to a binary value. How do we convert a regression model to a classification model? In order to map a logistic regression value to a binary category, you must define a classification threshold (also called the decision threshold). A value above that threshold indicates class A; a value below indicates class B. It is tempting to assume that the classification threshold should always be 0.5, but thresholds are problem-dependent, and are therefore values that you must tune. How can one determine a good decision threshold?
16. ### Classification: True vs False, Positive vs Negative A famous tale,

Consider, • ‘Wolf’ as a positive class • ‘No Wolf’ as a negative class
17. ### Classification: True vs False, Positive vs Negative Our "wolf-prediction" model,

there are four possible outcomes • A true positive is an outcome where the model correctly predicts the positive class. Similarly, a true negative is an outcome where the model correctly predicts the negative class. • A false positive is an outcome where the model incorrectly predicts the positive class. And a false negative is an outcome where the model incorrectly predicts the negative class.
18. ### Metrics for Classification Accuracy: Accuracy is the fraction of predictions

our model got right. Let’s take an example,
19. ### Metrics for Classification The accuracy of the model was 91%

which seems great!! Let’s take a closer look. • Of the 100 tumor examples, 91 are benign (90 TNs and 1 FP) and 9 are malignant (1 TP and 8 FNs). • Of the 91 benign tumors, the model correctly identifies 90 as benign. That's good. However, of the 9 malignant tumors, the model only correctly identifies 1 as malignant—a terrible outcome, as 8 out of 9 malignancies go undiagnosed! • In other words, our model is no better than one that has zero predictive ability to distinguish malignant tumors from benign tumors. Accuracy alone doesn't tell the full story when you're working with a class-imbalanced data set, like this one, where there is a significant disparity between the number of positive and negative labels. Hence we need better metrics.
20. ### Towards Better Metrics Precision: Precision attempts to answer: “Was the

model correct when it predicted a positive class?” Recall: Recall attempts to answer: “Out of all the possible positive class outcomes, how many did the model correctly identify?” Precision and recall are often in tension. That is, improving precision typically reduces recall and vice versa.
21. ### Precision and Recall We look at an email classification model.

Those to the right of the classification threshold are classified as "spam", while those to the left are classified as "not spam."

right.
23. ### Precision and Recall We shift the classification threshold to the

left. Hence we see that when precision increases recall decreases and vice versa.

25. ### ROC Curve An ROC curve (receiver operating characteristic curve) is

a graph showing the performance of a classification model at all classification thresholds. This curve plots two parameters- True Positive Rate (TPR) is a synonym for recall and is therefore defined as follows: False Positive Rate (FPR) is defined as follows: An ROC curve plots TPR vs. FPR at different classification thresholds. Lowering the classification threshold increases False Positives and True Positives, though typically not to the same degree.
26. ### Area Under ROC Curve (AUC) To compute the points in

an ROC curve, we could evaluate a logistic regression model many times with different classification thresholds, but this would be inefficient. Fortunately, there's an efficient, sorting-based algorithm that can provide this information for us, called AUC (Area under the ROC Curve).
27. ### AUC .AUC measures the entire two-dimensional area underneath the entire

ROC curve from (0,0) to (1,1). One way of interpreting AUC is as the probability that the model ranks a random positive example more highly than a random negative example. AUC ranges in value from 0 to 1. A model whose predictions are 100% wrong has an AUC of 0.0; one whose predictions are 100% correct has an AUC of 1.0.