FAI - ch5 - Remaining Topics
FAI - ch5 - Remaining Topics
FAI - ch5 - Remaining Topics
overfitting problem
The hypothesis is one of the commonly used concepts of statistics in Machine Learning. It is
specifically used in Supervised Machine learning, where an ML model learns a function that best
maps the input to corresponding outputs with the help of an available dataset.
Hypothesis space is defined as a set of all possible legal hypotheses; hence it is also known as a
hypothesis set. It is used by supervised machine learning algorithms to determine the best possible
hypothesis to describe the target function or best maps input to output.
It is often constrained by choice of the framing of the problem, the choice of model, and the choice
of model configuration.
The inductive bias (also known as learning bias) of a learning algorithm is the set of
assumptions that the learner uses to predict outputs of given inputs that it has not
encountered.
In machine learning, one aims to construct algorithms that are able to learn to predict a
certain target output.
To achieve this, the learning algorithm is presented some training examples that
demonstrate the intended relation of input and output values.
Then the learner is supposed to approximate the correct output, even for examples that
have not been shown during training.
Without any additional assumptions, this problem cannot be solved since unseen
situations might have an arbitrary output value.
The kind of necessary assumptions about the nature of the target function are subsumed
in the phrase inductive bias.
For Example:
Linear Regression algorithm assumes linearity, normality, and homoscedasticity (equal
variance)
Naive Bayes assumes that the data is Normally distributed, and conditional independence
exists between the independent features.
K-NN makes the assumption that the data points close together will be similar, hence the
new data point/unknown data will be placed with the majority of the neighbors.
Support Vector Machines assume that the margins should be large
There are several techniques to avoid overfitting in Machine Learning altogether listed below.
Collaborative Filtering
Ref: https://www.analyticsvidhya.com/blog/2022/02/introduction-to-collaborative-filtering/