- Dinesh Kumawat
- Aug 21, 2019
- Machine Learning
- Updated on: Oct 29, 2020

Logistic Regression is basically a predictive model analysis technique where the output (target) variables are discrete values for a given set of features or input (X). It is a very powerful yet simple supervised classification algorithm in machine learning.

Around 60% of the world’s classification problems can be solved by using the logistic regression algorithm. In this blog, we go all the basic concepts of logistic regression with code explanation.

Logistic regression is one of the **most common machine learning algorithms** used for binary classification. It predicts the probability of occurrence of a binary outcome using a logit function. It is a special case of linear regression as it predicts the probabilities of outcome using log function.

We use the activation function (sigmoid) to convert the outcome into categorical value. There are many examples where we can use logistic regression for example, it can be used for fraud detection, spam detection, cancer detection, etc.

Linear Regression is used when our dependent variable is continuous in nature for example weight, height, numbers, etc. and in contrast, Logistic Regression is used when the dependent variable is binary or limited for example: yes and no, true and false, 1 or 2 or 3 or 4, etc. Linear regression uses the ordinary least square method to minimize the error and arrives at the best possible solution, and the Logistic regression achieves the best outcomes by using the maximum likelihood method.

In the 19th century, people use linear regression on biology to predict health disease but it is very risky for example if a patient has cancer and its probability of malignant is 0.4 then in linear regression it will show that cancer is benign (because probability comes <0.5). That’s where Logistic Regression comes which only provides us with binary results.

It is a mathematical function having a characteristic that can take any real value and map it to between 0 to 1 shaped like the letter “S”. The sigmoid function also called a logistic function.

**Y = 1 / 1+e ^{-z}**

__Sigmoid function__

So, if the value of z goes to positive infinity then the predicted value of y will become 1 and if it goes to negative infinity then the predicted value of y will become 0. And if the outcome of the sigmoid function is more than 0.5 then we classify that label as class 1 or positive class and if it is less than 0.5 then we can classify it to negative class or label as class 0.

Sigmoid Function acts as an activation function in machine learning which is used to add non-linearity in a machine learning model, in simple words it decides which value to pass as output and what not to pass, there are mainly 7 types of Activation Functions which are used in machine learning and deep learning.

You can find the dataset here Dataset. First of all, before proceeding we first import all the libraries that we need to use in our algorithm

```
import numpy as np
import matplotlib.pyplot as plt
from matplotlib.colors import ListedColormap
import pandas as pd
from sklearn.model_selection import train_test_split
from sklearn.preprocessing import StandardScaler
from sklearn.metrics import accuracy_score
from sklearn.linear_model import LogisticRegression
```

After initializing all the libraries that we need in our algorithm know we have to import our dataset with the help of the pandas library and split our dataset into training and testing set with the help of the train_test_split library.

```
dataset = pd.read_csv(r'dataset.csv')
x = dataset.iloc[:,[2,3]].values
y = dataset.iloc[;, 4].values
x_train, x_test, y_train, y_test = train_test_split(x,y, test_size=0.2, random_state=0)
```

As we divide our dataset on the basis of train and test split know we have to scale our feature dataset with the help of StandardScaler library and apply logistic regression on the training set and check the accuracy sore with the help of accuracy_score library.

```
# Feature Scaling
sc = StandardScaler()
x_train = sc.fit_transform(x_train)
x_test = sc.transform(x_test)
# Fitting logistic regression to the training set
Classifier = LogisticRegression(random_state=0)
Classifier.fit(x_train, y_train)
# Predicting the test results
y_pred = classifier.predict(x_test)
print("Accuracy score: ", accuracy_score(y_test, y_pred))
# Accuracy score : 0.9125
```

We have successfully applied logistic regression on the training set and see that our accuracy scores come 89%. So, if we take on basis of algorithm it is not so much worse for prediction. As we get the accuracy score of our model now we can see a pictorial representation of our dataset first we have to visualize the result on the basis of the training dataset.

```
X_set, y set = X_test, y_test
X1, X2 = np.meshgrid(np.arange(start = X_set[:, 0].min()- 1 ,
stop = X_set[:, 0].max() + 1,
step = 0.01),
np.arange(start - X_set[:, 1].min() -1,
stop = X_set[:, 0].max() + 1,
step = 0.01)),
plt.contourf(x1, x2, classifier.predict
(np. array([X1.ravel(), X2.ravel()].T).reshape(X1.shape),
alpha = 0.75, Cmap = ListedColormap(('red', 'green')))
plt.xlim(X1.min(), X1.max())
plt.ylim(X1.min(), X1.max())
for i,j in enumerate(np.unique(y_set)):
plt.scatter(X_set[y_set== j, 0], X_set[y_set == j, 1],
c = ListedColormap(('red', 'green'))(i), label = j)
plt.title('Logistic Regression (Training set)')
plt.xlabel('Age')
plt.ylabel('Estimated Salary')
plt.legend()
plt.show()
```

__ Estimated Salary Result(training set )__

We plot a picture on the basis of age and estimated salary in which we bifurcate our result in a 0 and 1 value basis. In the same process, we apply for the test set and visualize our result how accurate our prediction is.

```
X_set, y set = X_test, y_test
X1, X2 = np.meshgrid(np.arange(start = X_set[:, 0].min()- 1 ,
stop = X_set[:, 0].max() + 1,
step = 0.01),
np.arange(start - X_set[:, 1].min() -1,
stop = X_set[:, 0].max() + 1,
step = 0.01)),
plt.contourf(x1, x2, classifier.predict
(np. array([X1.ravel(), X2.ravel()].T).reshape(X1.shape),
alpha = 0.75, Cmap = ListedColormap(('red', 'green')))
plt.xlim(X1.min(), X1.max())
plt.ylim(X1.min(), X1.max())
for i,j in enumerate(np.unique(y_set)):
plt.scatter(X_set[y_set== j, 0], X_set[y_set == j, 1],
c = ListedColormap(('red', 'green'))(i), label = j)
plt.title('Logistic Regression (Test set)')
plt.xlabel('Age')
plt.ylabel('Estimated Salary')
plt.legend()
plt.show()
```

__ Estimated Salary Result (Test Set)__

The main concept regarding this blog is to explain logistic regression and simple explanation via python code. I think the above blog is very helpful for you to clear your doubts regarding logistic regression more blogs are on the way to stay tuned with us! Keep exploring Analytics Steps.

0%

Reliance Jio and JioMart: Marketing Strategy, SWOT Analysis, and Working Ecosystem

READ MORE6 Major Branches of Artificial Intelligence (AI)

READ MORETop 10 Big Data Technologies in 2020

READ MOREWhat is the OpenAI GPT-3?

READ MOREIntroduction to Time Series Analysis: Time-Series Forecasting Machine learning Methods & Models

READ MORE7 types of regression techniques you should know in Machine Learning

READ MORE8 Most Popular Business Analysis Techniques used by Business Analyst

READ MOREHow Does Linear And Logistic Regression Work In Machine Learning?

READ MORE7 Types of Activation Functions in Neural Network

READ MOREHow is Artificial Intelligence (AI) Making TikTok Tick?

READ MORE

## Comments