GlobalLogic - Optimization Algorithms For Machine Learning
GlobalLogic - Optimization Algorithms For Machine Learning
In this use case scenario, we explore how an optimized machine learning model
Introduction
Employers generally consider attrition a loss of valuable employees and talent;
leave an organization, they take with them much-needed skills and qualifications they developed
during their tenure. There is no way for employers to know which employees will leave the
company, but a well-trained machine learning model can be used to predict attrition. We will
look at some of the optimization algorithms to improve the performance of the model.
Optimization is the most crucial part of machine learning algorithms. It begins with defining loss
function/cost function and ends with minimizing loss and cost using optimization algorithms
These help us maximize or minimize an error function. The internal parameters of a model play a
very important role in efficiently and effectively training a model and producing accurate results.
This is why we use various optimization algorithms to update and calculate appropriate and
optimum values of a model’s parameters. This, in turn, improves our model’s learning process, as
well as its output.
1) Dataset
Head out to *link* read about all the above topics in brief.
1)Batch Normalization
Batch normalization is a method used to normalize the inputs of each layer
the performance and stability of neural networks. This also makes more
and unit variance). This allows each layer to learn on a more stable distribution
We normalize the input layer by adjusting and scaling the activations, which
2) Grid-Search
depending on the type of model utilized. Grid-searching does not apply to only
one model type. Grid-searching can be applied to calculate the best parameters
to use for any given model across machine learning. It works in an iterative way.
For some of the parameters associated with the model, we enter good probable
values and the grid-search iterates through each of them, compares the result
for each value, and then gives you the parameters best suited for your model.
samples are selected randomly instead of using a whole data set for each
iteration or using data in the order they appear in the training set. We adjust the
total number of samples from a dataset used to calculate the gradient for
Clink on the link to read about the methods in detail. There are advantages and disadvantages
provided for each along with code snippets.
Conclusion
We implemented different models to predict attrition in a company, measured
However, in reality we might have many more data sets where optimization
model.
Finally, we have a working model to predict which employees will leave the
company and who will stay based on five input parameters with an accuracy of
almost 98 percent.