Logistic Regression

"The true logic of this world is the calculus of probabilities."
- James Clerk Maxwell, Scottish physicist

In the previous chapter, we took a look at using Ordinary Least Squares (OLS) to predict a quantitative outcome or, in other words, linear regression. It's now time to shift gears somewhat and examine how we can develop algorithms to predict qualitative outcomes. Such outcome variables could be binary (male versus female, purchase versus doesn't purchase, or a tumor is benign versus malignant) or multinomial categories (education level or eye color). Regardless of whether the outcome of interest is binary or multinomial, our task is to predict the probability of an observation belonging to a particular category of the outcome variable. In other words, we develop an algorithm to classify the observations.

To begin exploring classification problems, we'll discuss why applying the OLS linear regression isn't the correct technique and how the algorithms introduced in this chapter can solve these issues. We'll then look at the problem of predicting whether or not a banking customer is satisfied. To tackle this problem, we'll begin by building and interpreting a logistic regression model. We'll also start examining a univariate method to select features. Next, we'll turn to multivariate regression splines and discover ways to choose the best overall algorithm. This chapter will set the stage for more advanced machine learning methods in subsequent chapters.

We'll be covering the following topics in this chapter:

  • Classification methods and linear regression
  • Logistic regression
  • Model training and evaluation
..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
3.129.26.22