Machine Learning
2150602
DR. RAHUL DUBEY
ASSISTANT PROFESSOR
DEPARTMENT OF CSE
Syllabus
Unit – I
Introduction to Machine Learning: Learning, Traditional Vs Machine Learning,
Types Of ML, Classification and Regression model, Challenges faced by ML, Steps
of developing an ML model, Bias and Variance, Underfitting and Overfitting,
Regularization, Data visualization, Outlier, Testing and validating, K cross
Dr. Rahul Dubey
validation, Hyperparameter tuning, Model Selection.
Unit – II:
Model optimization and Evaluation: Parametric and non- Parametric model, Learner
performance evaluation, confusion matrix, Recall, accuracy, precision, Model
optimization, Cost/Loss Function, Derivative of cost function and non-derivative
cost function, Gradient descent, Mini-batch Gradient Descent (sckit-learn),
Stochastic Gradient descent(sckit-learn), Momentum(sckit-learn).
2
Unit – III:
Supervised Machine Learning Algorithm with python: Model Complexity vs Dataset
Size,
Supervised Machine Learning Algorithms, k-Nearest Neighbors, Linear Regression, RMSE,
Logistic Regression, Log Loss, Support Vector Machine, Hinge Loss, Kernel Trick,
polynomial
Kernal, Decision Trees, Gini impurity.
Dr. Rahul Dubey
Unit – IV:
Ensemble Learner with python: Ensemble learner, Bagging, Pasting , Voting Classifiers,
Out of-Bag, Evaluation, Random Patches and Random Subspaces , Random Forests , Extra-
Trees, Boosting , AdaBoost, Gradient Boosting, Stacking
Unit –V:
Unsupervised Machine Learning with python: The Curse of Dimensionality, Principal
component analysis, Clustering, K-Means, Limits of K-Means, Clustering, DBSCAN.
ElasticNet, Transformer-based models, Explainable ML, Ethical ML, AutoML, Capstone
Project and Case Studies: Build an end-to-end supervised or unsupervised ML pipeline.
3
Text Books
1) Tom M. Mitchell, “Machine Learning”, McGraw-Hill Education
(INDIAN EDITION), 2013
2) Christopher M. Bishop, "Pattern Recognition and Machine Learning", by
Springer, 2007
3)
Dr. Rahul Dubey
Amanda Casari, Alice Zheng, “Feature Engineering for Machine
Learning”, O’Reilly, 2018.
4) Andreas Muller, "Introduction to Machine Learning with Python: A
Guide for Data Scientists", Shroff/O'Reilly; First edition (2016)
4
Artificial Intelligence, Machine
Learning and Deep Learning & Data
Science
Dr. Rahul Dubey
5
What Is Machine Learning?
Machine Learning is the science (and art) of
programming computers so they can learn from data.
OR
Dr. Rahul Dubey
[Machine Learning is the] field of study that gives
computers the ability to learn without being explicitly
programmed.
Arthur Samuel, 1959
What Is Machine Learning?
More engineering-oriented one:
A computer program is said to learn from experience E
with respect to some task T and some performance
Dr. Rahul Dubey
measure P, if its performance on T, as measured by P,
improves with experience E.
Tom Mitchell, 1997
Analogy
spam filter is a Machine Learning program that can learn to
flag spam given examples of spam emails (e.g., flagged by
users) and examples of regular (no spam, also called
Dr. Rahul Dubey
“ham”) emails
In this case, the task T is to flag spam for new emails, the
experience E is the training data, and the performance
measure P needs to be defined; for
example, you can use the ratio of correctly classified emails.
This particular performance measure is called accuracy and
it is often used in classification tasks.
Traditional Programming
Data
Computer Output
Program
Dr. Rahul
Machine Learning
Dubey
Data
Computer Program
Output
Traditional Programming
Dr. Rahul Dubey
Machine Learning Approach
Dr. Rahul Dubey
To summarize, Machine Learning is great for:
Problems for which existing solutions require a lot of
hand tuning or long lists of rules: one Machine
Learning algorithm can often simplify code and
Dr. Rahul Dubey
perform better.
Complex problems for which there is no good solution
at all using a traditional approach: the best Machine
Learning techniques can find a solution.
To summarize, Machine Learning is great for:
Fluctuating environments: a Machine Learning system
can adapt to new data.
Getting insights about complex problems and large
Dr. Rahul Dubey
amounts of data.
ML, DL History
first machine learning • In 1957, Rosenblatt
system was developed in proposed the
the 1950s perceptron
In 1952, Arthur Samuel • It was good linear
developedDr. Rahul
a program for Dubey
classifier
playing Checkers it was • But in 1962 minsky
learning Term coined told it can not model
“Machine learning” all problems
• Then no work till 1980’s
ML, DL History
1986 , decision tree • 1981 ,resurgence of
learning, specifically the neural network by
ID3 algorithm
In the 90s, machine multilayer perceptron
learning embraced and back propagation
Dr. Rahul
statistics to a large extent
support vector machines
Dubey
algorithm
were proposed • Back propagation is
In 1997 Another strong key ingredient of
approach today’s neural
ensembles or boosting network
Eg. Adaboost
ML, DL History
kernalized version of • Only Hardware
SVM was proposed acceleration was
near 2000s going on
Convex optimization
Dr. Rahul Dubey
approach is proposed
“Big Mile Stone”
New Ensemble
Random forest in 2001
ML, DL History
• new era in neural network
called Deep Learning get
popular in 2005 with
Hinton's lecue and Andrew
research work
• In 1994, the first self driving
Dr. Rahul Dubey
car made a road test
• in 1997, Deep
Blue beat the world
champion Gary Kasparov in
the game of chess
• in 2009 we have
Google building self driving
cars
ML, DL History
• Watson, again from IBM,
won the popular game of
Jeopardy; 2014
• In 2015 CNN surpasses the
Dr. Rahul Dubey
human vision system
• In 2015 machine translation
systems
What Next ?
• How to automate the design of Deep learning
architectures
• How to make the things interpretable
Dr. Rahul Dubey
• Explainable AI or XAI or Interpretable AI
Machine Learning Cycle
Dr. Rahul Dubey