Data Science
Data Science
Regression estimates the relationship between the target and the independent
variable.
It is used to find the trends in data.
It helps to predict real/continuous values.
By performing the regression, we can confidently determine the most
important factor, the least important factor, and how each factor is affecting
the other factors.
Types of Regression
There are various types of regressions which are used in data science and machine
learning. Each type has its own importance on different scenarios, but at the core,
all the regression methods analyze the effect of the independent variable on
dependent variables.
Here we are discussing some important types of regression which are given below:
Linear Regression
Logistic Regression
Polynomial Regression
Support Vector Regression
Decision Tree Regression
Random Forest Regression
Ridge Regression
Lasso Regression.
Linear Regression:
Linear regression is a statistical regression method which is used for
predictive analysis.
It is one of the very simple and easy algorithms which works on regression
and shows the relationship between the continuous variables.
It is used for solving the regression problem in machine learning.
Below is the mathematical equation for Linear regression:
Y= aX+b
Here, Y = dependent variables (target variables),
X= Independent variables (predictor variables),
A and b are the linear coefficients
Logistic Regression:
Logistic regression is another supervised learning algorithm which is used to solve
the classification problems. In classification problems, we have dependent variables
in a binary or discrete format such as 0 or 1.
Logistic regression algorithm works with the categorical variable such as 0 or 1, Yes
or No, True or False, Spam or not spam, etc.
There are three types of logistic regression:
Binary(0/1, pass/fail)
Multi(cats, dogs, lions)
Ordinal(low, medium, high)
Polynomial Regression:
Polynomial Regression is a type of regression which models the non-linear dataset
using a linear model.
It is similar to multiple linear regression, but it fits a non-linear curve between the
value of x and corresponding conditional values of y.
Suppose there is a dataset which consists of datapoints which are present in a non-
linear fashion, so for such case, linear regression will not best fit to those
datapoints. To cover such datapoints, we need Polynomial regression.
Ridge Regression:
Ridge regression is one of the most robust versions of linear regression in which a
small amount of bias is introduced so that we can get better long term predictions.
The amount of bias added to the model is known as Ridge Regression penalty. We
can compute this penalty term by multiplying with the lambda to the squared weight
of each individual features.
Lasso Regression:
Lasso regression is another regularization technique to reduce the complexity of the
model.
It is similar to the Ridge Regression except that penalty term contains only the
absolute weights instead of a square of weights.