Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Ai ML

Download as pdf or txt
Download as pdf or txt
You are on page 1of 2

5 3 3 3 3 3 - - - - - - - 3 3 2

C 3 3 2.6 2.2 2.2 - - - - - - - 2.8 2.2 1.4


1 - low, 2 - medium, 3 - high, ‘-' - no correlation

CS3491 ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING LTP C


3 02 4
COURSE OBJECTIVES:
The main objectives of this course are to:
 Study about uninformed and Heuristic search techniques.
 Learn techniques for reasoning under uncertainty
 Introduce Machine Learning and supervised learning algorithms
 Study about ensembling and unsupervised learning algorithms
 Learn the basics of deep learning using neural networks

UNIT I PROBLEM SOLVING 9


Introduction to AI - AI Applications - Problem solving agents – search algorithms – uninformed
search strategies – Heuristic search strategies – Local search and optimization problems –
adversarial search – constraint satisfaction problems (CSP)

UNIT II PROBABILISTIC REASONING 9


Acting under uncertainty – Bayesian inference – naïve bayes models. Probabilistic reasoning –
Bayesian networks – exact inference in BN – approximate inference in BN – causal networks.

UNIT III SUPERVISED LEARNING 9


Introduction to machine learning – Linear Regression Models: Least squares, single & multiple
variables, Bayesian linear regression, gradient descent, Linear Classification Models: Discriminant
function – Probabilistic discriminative model - Logistic regression, Probabilistic generative model –
Naive Bayes, Maximum margin classifier – Support vector machine, Decision Tree, Random forests

UNIT IV ENSEMBLE TECHNIQUES AND UNSUPERVISED LEARNING 9


Combining multiple learners: Model combination schemes, Voting, Ensemble Learning - bagging,
boosting, stacking, Unsupervised learning: K-means, Instance Based Learning: KNN, Gaussian
mixture models and Expectation maximization

UNIT V NEURAL NETWORKS 9


Perceptron - Multilayer perceptron, activation functions, network training – gradient descent
optimization – stochastic gradient descent, error backpropagation, from shallow networks to deep
networks –Unit saturation (aka the vanishing gradient problem) – ReLU, hyperparameter tuning,
batch normalization, regularization, dropout.
45 PERIODS
PRACTICAL EXERCISES: 30 PERIODS
1. Implementation of Uninformed search algorithms (BFS, DFS)
2. Implementation of Informed search algorithms (A*, memory-bounded A*)
3. Implement naïve Bayes models
4. Implement Bayesian Networks
5. Build Regression models
6. Build decision trees and random forests

98
7. Build SVM models
8. Implement ensembling techniques
1. Implement clustering algorithms
2. Implement EM for Bayesian networks
3. Build simple NN models
4. Build deep learning NN models

OUTCOMES:
At the end of this course, the students will be able to:
CO1: Use appropriate search algorithms for problem solving
CO2: Apply reasoning under uncertainty
CO3: Build supervised learning models
CO4: Build ensembling and unsupervised models
CO5: Build deep learning neural network models
TOTAL:75 PERIODS

TEXT BOOKS:
1. Stuart Russell and Peter Norvig, “Artificial Intelligence – A Modern Approach”, Fourth
Edition, Pearson Education, 2021.
2. Ethem Alpaydin, “Introduction to Machine Learning”, MIT Press, Fourth Edition, 2020.

REFERENCES
1. Dan W. Patterson, “Introduction to AI and ES”, Pearson Education,2007
2. Kevin Night, Elaine Rich, and Nair B., “Artificial Intelligence”, McGraw Hill, 2008
3. Patrick H. Winston, "Artificial Intelligence", Third Edition, Pearson Education, 2006
4. Deepak Khemani, “Artificial Intelligence”, Tata McGraw Hill Education, 2013
(http://nptel.ac.in/)
5. Christopher M. Bishop, “Pattern Recognition and Machine Learning”, Springer, 2006.
6. Tom Mitchell, “Machine Learning”, McGraw Hill, 3rd Edition,1997.
7. Charu C. Aggarwal, “Data Classification Algorithms and Applications”, CRC Press, 2014
8. Mehryar Mohri, Afshin Rostamizadeh, Ameet Talwalkar, “Foundations of Machine
Learning”, MIT Press, 2012.
9. Ian Goodfellow, Yoshua Bengio, Aaron Courville, “Deep Learning”, MIT Press, 2016
CO’s-PO’s & PSO’s MAPPING
CO PO1 PO2 PO3 PO4 PO5 PO6 PO7 PO8 PO9 PO10 PO11 PO12 PSO1 PSO2 PSO3
1 3 2 2 3 1 3 2 - - - - 1 3 3 3
2 3 2 2 3 1 3 2 - - - - 1 3 3 3
3 1 2 1 3 2 3 2 - - - - 1 3 3 3
4 1 2 3 1 3 3 2 - - - - 1 3 3 3
5 2 2 2 - 3 3 2 - - - - 1 3 3 3
CO 2 2 2 2 2 3 2 - - - - 1 3 3 3

1 - low, 2 - medium, 3 - high, ‘-' - no correlation

99

You might also like