KEMBAR78
Reuqired ppt for machine learning algirthms and part | PPTX
Quick Insights about Machine learning, Neural
Network Algorithms, & AR Model
Siddhesh Mhatre
 Review Papers
 Predictive analysis using machine learning: Review of trends and methods
 Machine learning and deep learning based predictive quality in manufacturing: a systematic review
 Healthcare predictive analytics using machine learning and deep learning techniques: a survey
 What is Machine learning?
 It is a subset of Artificial Intelligence.
 The ability of Machines to solve any kind of problem without being explicitly programmed.
 Types of Machine learning: -
 Supervised learning
 Unsupervised learning
 Semi-supervised learning
 Reinforcement learning
 Supervised learning: -
 Supervised learning is basically that kind of learning which we can used for labelled data or previously
available data. (i.e. Historical data)
 And, It is further divided into two classes:-
1. Classification (Finds the relation between discrete input & output)
2. Regression (Finds the relation between two or more independent variables)
 Decision Trees
 This technique is most powerful technique which used for classification.
 It is a flowchart-like tree structure where each internal node denotes the feature, branches denote
the rules, and the leaf nodes denote the result of the algorithm.
 Example
 Decision Tree can be Interpreted in following ways: -
 Visualization
 Tree based
 Program based
 Visualization
 Tree based
 Program based
 Preliminary Concepts used in Decision Tree are : -
 Entropy
 Gini Impurity
 Information Gain
 Entropy
 Entropy is the measure of the degree of randomness or uncertainty in the dataset. In the
case of classifications, It measures the randomness based on the distribution of class
labels in the dataset.
 It lies from 0 to 1 as 0 we can say the data is homogeneous. When its goes to 1, we can
say there are lots of uncertainties in a data. (Multiple classes).
 It is formulated as: -
 Gini Impurity
 Gini Impurity is a score that evaluates how accurate a split is among the classified groups.
 It is also lies from 0 to 0.5. 0 as observations belong to same class and 0.5 is random
distribution of elements with classes. So less Gini impurity score is highly recommended.
 It is formulated as: -
Purpose of Entropy and
Gini Impurity is same
Only difference is Gini
Impurity has max value
0.5 and Entropy has max
value 1.0
 Information Gain
 Information gain measures the reduction in entropy or variance that results from
splitting a dataset based on a specific property.
 The higher the information gain, the more valuable the feature is in predicting the
target variable.
 It is formulated by,
 Hyperparameters
Hyperparameters coming under Decision tree are: -
 Pure node
 Maximum number of leaf nodes
 Minimum Information gain
 Maximum Depth
 Minimum samples at leaf node
 Minimum impurity to split
 Random Forest
 This technique is used for both classification
and regression.
 This technique is based on a concept of
ensemble learning, which is a process of
combining multiple classifiers to solve a
complex problem and to improve the
performance of the model.
 Ensemble models
• Ensemble – Group of things Combine different ML models to create a more powerful
model.
Following Ensemble models are: -
1. Bagging
2. Boosting
3. Stacking
4. Cascading
 Bagging – Bootstrap Aggregation
• Bootstrap – Sampling with replacement
• Aggregation – Combining
• Random forest is a Bagging based ensemble model
• Random forest can be used for Regression / Classification
 Bagging:- Process
 Bagging (In case of classification)
 Bagging (In case of regression)
 Hyperparameters
Hyperparameters coming under Decision tree are: -
 Number of base learners
 Decision Tree hyperparameters
 Why Random forest?
 No Overfitting
1. Use of Multiple trees to reduce the use of overfitting
2. Training time is less
 High Accuracy
1. Runs efficiently on large database
2. For large data, it produces highly accurate predictions.
 Estimates Missing data
1. Random forest can maintain accuracy when a large proportion of data is missing
 Naïve Bayes Classifier
 It is a classification technique based on Bayes’ Theorem with an assumption of
independence among predictors.
 In simple terms, a Naive Bayes classifier assumes that the presence of a
particular feature in a class is unrelated to the presence of any other feature.
 Despite its simplicity, it often performs surprisingly well in practice and serves as
a baseline model for many classification tasks.
 Bayesian Method
 Convert the data set into a frequency table
 Create Likelihood table by finding the probabilities
 Now, use Naive Bayesian equation to calculate the posterior probability for each class.
 The class with the highest posterior probability is the outcome of prediction.
 Types of Naive Bayes
1. Gaussian Naïve Bayes: -
• Variables are continuous in nature.
• It assumes that all the variables have a normal distribution
2. Multinominal Naïve Bayes: -
• The features represent the frequency.
• If you have a text document, and you extract all the unique words and create multiple features
where each feature represents the count of the word in the document.
3. Bernoulli Naïve Bayes: -
• This is used when features are binary.
Pros:
 It is easy and fast to predict class of test data set. It also perform well in multi class
prediction.
 When assumption of independence holds, a Naive Bayes classifier performs better
compared to other models like logistic regression and you need less training data.
 It perform well in case of categorical input variables compared to numerical
variable(s).
Cons:
If categorical variable has a category (in test data set), which was not observed in training data set,
then model will assign a 0 (zero) probability and will be unable to make a prediction. This is often known
as “Zero Frequency”. To solve this, we can use the smoothing technique. One of the simplest smoothing
techniques is called Laplace estimation.
On the other side naive Bayes is also known as a bad estimator, so the probability outputs
from predict_proba are not to be taken too seriously.
Another limitation of Naive Bayes is the assumption of independent predictors. In real life, it is almost
impossible that we get a set of predictors which are completely independent.
 Logistic Regression
 Definition of logistic regression: Logistic regression is a
statistical technique used to model the probability of a
binary outcome based on one or more predictor variables.
 logistic (sigmoid) function: The logistic function, also
known as the sigmoid function, maps any real-valued input
to a value between 0 and 1. It is the core component of
logistic regression and helps transform the output into
probabilities.
 Logistic function
Mathematical representation of the logistic function: The logistic function is represented as
where Z is the linear combination of input features and their corresponding coefficients.
Graphical representation of the logistic curve: A graphical representation of the logistic curve
shows an S-shaped curve that asymptotically approaches 0 and 1 on the y-axis.
 Advantages and Limitations
 Advantages of logistic regression: Logistic regression offers a simple and interpretable model,
provides probabilities as outputs, and can handle both linear and nonlinear relationships.
 Limitations of logistic regression: It assumes a linear relationship between features and log-odds, is
vulnerable to overfitting with a large number of features and is limited to binary classification.
 K-Nearest Neighbor (KNN)
 The K-Nearest Neighbors (KNN) algorithm is a
supervised machine learning method employed to
tackle classification and regression problems.
 The KNN algorithm works by finding the K nearest
neighbors to a given data point based on a distance
metric, such as Euclidean distance. The class or
value of the data point is then determined by the
majority vote or average of the K neighbors. This
approach allows the algorithm to adapt to different
patterns and make predictions based on the local
structure of the data.
 Distance Metrics Used in KNN Algorithm
We use following distance matrices for following algorithms:
1. Euclidean Distance: Euclidean distance can also be visualized as the length of the straight line that joins the two
points which are into consideration.
2. Manhattan Distance: This metric is calculated by summing the absolute difference between the coordinates
of the points in n-dimensions.
3. Minkowski Distance
We can say that the Euclidean, as well as the Manhattan distance, are special cases of the Minkowski
distance.
From the formula above we can say that when p = 2 then it is the same as the formula for the Euclidean
distance and when p = 1 then we obtain the formula for the Manhattan distance.
 Steps for KNN algorithm
1. Selecting the optimal value of K
2. Calculating distance
3. Finding Nearest Neighbors
4. Voting for Classification or
Taking Average for Regression
 Easy to implement - as the complexity of the algorithm is not that high.
 Adapts Easily – As per the working of the KNN algorithm it stores all the data in memory storage and
hence whenever a new example or data point is added then the algorithm adjusts itself as per that new
example and has its contribution to the future predictions as well.
 Few Hyperparameters – The only parameters which are required in the training of a KNN algorithm
are the value of k and the choice of the distance metric which we would like to choose from our
evaluation metric.
 Advantages of the KNN Algorithm
 Does not scale – As we have heard about this that the KNN algorithm is also considered a Lazy Algorithm. The main
significance of this term is that this takes lots of computing power as well as data storage. This makes this algorithm both
time-consuming and resource exhausting.
 Curse of Dimensionality – There is a term known as the peaking phenomenon according to this the KNN algorithm is
affected by the curse of dimensionality which implies the algorithm faces a hard time classifying the data points properly
when the dimensionality is too high.
 Prone to Overfitting – As the algorithm is affected due to the curse of dimensionality it is prone to the problem of
overfitting as well. Hence generally feature selection as well as dimensionality reduction techniques are applied to deal
with this problem.
 Disadvantages of the KNN Algorithm
 Model Evaluation Metrics
 Precision: -
Precision attempts to answer the following question:
What proportion of positive identifications was actually correct?
Precision is defined as follows: -
 Recall: -
Recall attempts to answer the following question:
What proportion of actual positives was identified correctly?
Mathematically, recall is defined as follows:
The F1-score is a measure of a model’s performance that combines precision and recall. It is
defined as the harmonic mean of precision and recall, where the best value is 1 and the worst
value is 0.
 F1-Score
 Accuracy
We can find accuracy by calculating the ratio of correct predictions to the total number of input Samples.
 Sensitivity and Specificity
Count up the total number of each label (TP, FP, TN, FN) over a large dataset. In ROC analysis, we use two
statistics:
 Sensitivity: - Can be thought of as the likelihood of spotting a positive case when presented with one.
 Specificity: - Can be thought of as the likelihood of spotting a negative case when presented with one.
 ROC Curve
An ROC curve (receiver operating characteristic curve) is a graph showing the performance of a
classification model at all classification thresholds.
 Unsupervised learning: -
 Unsupervised learning, unlike supervised learning, does not come with labels (no output vectors).
 The objective of unsupervised learning is to analyze the structure of the data and extract useful
information from it without any explicit indication of the expected result.
 Unsupervised learning includes two sub-types:
1. Clustering
2. Dimensionality reduction.
 K-means clustering
 K-means is a partitioning (clustering) method that
divides a dataset into ‘k’ distinct, non-overlapping
subsets (clusters) based on similarity, aiming to
minimize the variance within each cluster.
 It is one of the most popular unsupervised learning
algorithms that deals with well-known clustering
problem.
 Since it can handle large sized dataset, It is the most
computationally efficient than any other clustering
methods.
 How k-means clustering works?
 We are given a data set of items, with certain features, and values for these features (like a vector).
The task is to categorize those items into groups. To achieve this, we will use the K-means
algorithm, an unsupervised learning algorithm. ‘K’ in the name of the algorithm represents the
number of groups/clusters we want to classify our items into.
 The algorithm works as follows:
First, we initialize the k points, called means or cluster centroids.
We categorize each item to its closest mean, and we update the mean’s coordinates, which are the averages of
the items categorized in that cluster so far.
We repeat the process for a given number of iterations and at the end, we have our clusters.
 How do we choose the right k value?
 The Elbow Method is a technique used to determine
the optimal number of clusters (k) in the k-means
clustering algorithm.
 What does Elbow Method do?
We calculate the Within-Cluster Sum of Squares (WCSS) for different values of k.
WCSS measures the sum of square distances between the centroids and each data point.
As k increases, WCSS decreases because clusters become more compact.
However, beyond a certain point, the improvement becomes marginal.
 Graphical Approach:
We plot k against its corresponding WCSS value.
The graph often resembles an “elbow.”
The optimal k value is where the graph starts to look like a straight line.
From the above graph, we can observe that at k=2 and k=3 elbow-like situation. So, we are
considering K=3
 Principal Component Analysis (PCA)
 In modern data analysis, principal component analysis (PCA) is an essential tool as it provides a guide
for extracting the most information from a dataset, compressing the data size by keeping only those
important features without losing much information, and simplifying the description of a dataset.
 It works on the condition that while the data in a higher dimensional space is mapped to data in a lower
dimension space, the variance of the data in the lower dimensional space should be maximum.
 The main goal of Principal Component Analysis (PCA) is to reduce the dimensionality of a dataset while
preserving the most important patterns or relationships between the variables without any prior
knowledge of the target variables.
Before PCA & After PCA
 Principal Component Analysis (PCA) is a technique for dimensionality reduction that
identifies a set of orthogonal axes, called principal components, that capture the maximum
variance in the data. The principal components are linear combinations of the original
variables in the dataset and are ordered in decreasing order of importance. The total variance
captured by all the principal components is equal to the total variance in the original dataset.
 The first principal component captures the most variation in the data, but the second
principal component captures the maximum variance that is orthogonal to the first principal
component, and so on.
 Principal Component Analysis can be used for a variety of purposes, including data
visualization, feature selection, and data compression. In data visualization, PCA can be
used to plot high-dimensional data in two or three dimensions, making it easier to interpret.
 Mechanism of PCA
 In feature selection, PCA can be used to identify the most important variables in a
dataset. In data compression, PCA can be used to reduce the size of a dataset without
losing important information.
 In Principal Component Analysis, it is assumed that the information is carried in the
variance of the features, that is, the higher the variation in a feature, the more information
that features carries.
Overall, PCA is a powerful tool for data analysis and can help to simplify complex datasets,
making them easier to understand and work with.
Neural Networks
 Neural Networks are computational models
that mimic the complex functions of the
human brain. The neural networks consist of
interconnected nodes or neurons that process
and learn from data, enabling tasks such as
pattern recognition and decision making in
machine learning.
 It is composed of an input layer, one or more
hidden layers, and an output layer made up of
layers of artificial neurons that are coupled.
 Types of Neural Networks: -
• Feedforward Networks
• Multilayer Perceptron (MLP)
• Convolutional Neural Network (CNN)
• Recurrent Neural Network (RNN)
• Long Short-Term Memory (LSTM)
 Feedforward Networks
 A feedforward neural network is a type of artificial neural network in which nodes’ connections do not
form a loop.
 Often referred to as a multi-layered network of neurons, feedforward neural networks are so named
because all information flows in a forward manner only.
 The data enters the input nodes, travels through the hidden layers, and eventually exits the output nodes.
The network is devoid of links that would allow the information exiting the output node to be sent back
into the network.
 The purpose of feedforward neural networks is to approximate functions.
Working of Feed-forward Neural Network
 Components of Feedforward neural network
 Layer of Input
 Hidden Layer
 Output layer
 Neurons Weights
 Multilayer Perceptron (MLP)
 A Multilayer Perceptron (MLP) is a type of feedforward artificial neural network. It consists of fully
connected neurons with a nonlinear kind of activation function, organized in at least three layers
 Every node in the multi-layer perception uses a sigmoid activation function. The sigmoid activation
function takes real values as input and converts them to numbers between 0 and 1 using the sigmoid
formula.
Schematic Representation of MLP
 Multilayer Perceptron falls under the category of feedforward algorithms, because inputs are
combined with the initial weights in a weighted sum and subjected to the activation function, just
like in the Perceptron. But the difference is that each linear combination is propagated to the next
layer.
 Multi-layer perceptrons (MLPs) have a wide range of applications in various fields like, in Image
classification, Speech recognition, Natural language Processing, Finance, Computer vision,
Machine translation etc.
 Convolutional Neural Network (CNN)
 CNNs are special types of neural networks
inspired by human visual cortex used in
computer vision.
 It is automatic feed-forward neural
network in which information transfers
exclusively in the forward direction.
 A typical CNN architecture consists three
types of layers: convolution, pooling, and
classification.
 Here, Even-numbered layers are used for
convolutions, while odd numbered layers
used for pooling operations.
 The Convolutional layer applies filters to the input image to extract features, the Pooling layer
downsamples the image to reduce computation, and the fully connected layer makes the final
prediction. The network learns the optimal filters through backpropagation and gradient descent.
Advantages of Convolutional Neural Networks (CNNs):
 Good at detecting patterns and features in images, videos, and audio signals.
 Robust to translation, rotation, and scaling invariance.
 End-to-end training, no need for manual feature extraction.
 Can handle large amounts of data and achieve high accuracy.
Disadvantages of Convolutional Neural Networks (CNNs):
 Computationally expensive to train and require a lot of memory.
 Can be prone to overfitting if not enough data or proper regularization is used.
 Requires large amounts of labeled data.
 Interpretability is limited, it’s hard to understand what the network has learned.
 Recurrent Neural Network (RNN)
 Recurrent Neural Network (RNN) is a type of Neural Network where the output from the previous
step is fed as input to the current step.
 In traditional neural networks, all the inputs and outputs are independent of each other. Still, in cases
when it is required to predict the next word of a sentence, the previous words are required and hence
there is a need to remember the previous words. Thus RNN came into existence, which solved this
issue with the help of a Hidden Layer.
 The main and most important feature of RNN is its Hidden state, which remembers some information
about a sequence. The state is also referred to as Memory State since it remembers the previous input
to the network. It uses the same parameters for each input as it performs the same task on all the
inputs or hidden layers to produce the output. This reduces the complexity of parameters, unlike other
neural networks.
Mechanism of RNN
 Types Of RNN
There are four types of RNNs based on the
number of inputs and outputs in the network.
 One to One
 One to Many
 Many to One
 Many to Many
 How does RNN work?
The Recurrent Neural Network consists of multiple fixed activation function units, one for
each time step. Each unit has an internal state which is called the hidden state of the unit. This
hidden state signifies the past knowledge that the network currently holds at a given time
step. This hidden state is updated at every time step to signify the change in the knowledge of
the network about the past.
 Applications of Recurrent Neural Network
• Language Modelling and Generating Text
• Speech Recognition
• Machine Translation
• Image Recognition, Face detection
• Time series Forecasting
 Long Short-Term Memory (LSTM)
 LSTM networks are an extension of recurrent neural networks (RNN’s) mainly introduced to
handle situations where RNNs fail.
 It fails to store information for a longer period of time. At times, a reference to certain
information stored quite a long time ago is required to predict the current output. But RNNs
are absolutely incapable of handling such “long-term dependencies”.
 There is no finer control over which part of the context needs to be carried forward and how
much of the past needs to be ‘forgotten’.
 Other issues with RNNs are exploding and vanishing gradients which occur during the
training process of a network through backtracking.
Mechanism of Different LSTM’S
 Figure-A represents what a basic LSTM network looks like. Only one layer of LSTM between an input
and output layer has been shown here.
 Figure B represents Deep LSTM which includes a number of LSTM layers in between the input and
output. The advantage is that the input values fed to the network not only go through several LSTM
layers but also propagate through time within one LSTM cell. Hence, parameters are well distributed
within multiple layers. This results in a thorough process of inputs in each time step.
 Figure C represents LSTM with the Recurrent Projection layer where the recurrent connections are taken
from the projection layer to the LSTM layer input. This architecture was designed to reduce the high
learning computational complexity (O(N)) for each time step) of the standard LSTM RNN.
 Figure D represents Deep LSTM with a Recurrent Projection Layer consisting of multiple LSTM layers
where each layer has its own projection layer. The increased depth is quite useful in the case where the
memory size is too large. Having increased depth prevents overfitting in models as the inputs to the network
need to go through many nonlinear functions.
 Applications of LSTM Networks
Language Modelling
Image processing
Speech and Handwriting Recognition
Music generation
Language Translation
 Drawback of LSTM Networks
Problem of vanishing gradients
Sometimes Inefficient
Different random weight initialization
More expectations required
Time Series
 Series of observations (data-points) recorded
over time-intervals (usually equal time-
intervals)
 Note that in practice, data points are always
noted at discrete times → time-series is
NOT continuous
Components of time-series
Trend:- Long term behaviour
Seasonal:- variations short term periodic tendency
Cyclic variations:- repeating oscillatory tendency
Random, irregular, Noise, residual:-unpredictable instances
 Auto Regressive models (AR Model)
Self
Relation between two variables
predicting value of one variable to another variable
“So, Basically AR model is Predict future values from past self values”
The equation of AR model is given by :-
Where,
β is Coefficients,
Xt is targeted/Predicted value
Xt-k is Pervious values
εt is error at time t
 What is the effect of AR model?
 Autocorrelation function (ACF):-
 Gives Direct as well as Indirect effect.
 We estimate ACF by using Person’s Correlation
Coefficient.
 The equation of ACF for AR model is
 Partial-Autocorrelation function (PACF):-
 Gives Direct effect only.
 We estimate PACF by using Yule-walker’s
Equation.
 The equation of PACF for AR model is
 Applications of Auto-Regressive model
Benefits of Autoregressive Models:
 Simplicity: AR models are relatively simple to understand and implement. They rely on past values of
the time series to predict future values, making them conceptually straightforward.
 Interpretability: The coefficients in an AR model have clear interpretations. They represent the
strength and direction of the relationship between past and future values, making it easier to derive
insights from the model.
 Useful for Stationary Data: AR models work well with stationary time series data. Stationary data
have stable statistical properties over time, which is an assumption that AR models are built upon.
 Efficiency: AR models can be computationally efficient, especially for short time series or when you
have a reasonable amount of data.
Drawbacks of Autoregressive Models:
 Stationarity Assumption: AR models assume that the time series is stationary, meaning that its statistical
properties do not change over time. In practice, many real-world time series are non-stationary, requiring
preprocessing steps like differencing.
 Limited to Short-Term Dependencies: AR models are not well-suited for capturing long-term dependencies
in data. They are primarily designed for modeling short-term temporal patterns.
 Lag Selection: Choosing the appropriate lag order (p) in an AR model can be challenging. Selecting too few
lags may lead to underfitting, while selecting too many may lead to overfitting. Techniques like ACF and
PACF plots are used to determine the lag order.
 Sensitivity to Noise: AR models can be sensitive to random noise in the data. This sensitivity can lead to
overfitting, especially when dealing with noisy or irregular time series.
Thank you

Reuqired ppt for machine learning algirthms and part

  • 1.
    Quick Insights aboutMachine learning, Neural Network Algorithms, & AR Model Siddhesh Mhatre
  • 2.
     Review Papers Predictive analysis using machine learning: Review of trends and methods  Machine learning and deep learning based predictive quality in manufacturing: a systematic review  Healthcare predictive analytics using machine learning and deep learning techniques: a survey
  • 3.
     What isMachine learning?  It is a subset of Artificial Intelligence.  The ability of Machines to solve any kind of problem without being explicitly programmed.  Types of Machine learning: -  Supervised learning  Unsupervised learning  Semi-supervised learning  Reinforcement learning
  • 4.
     Supervised learning:-  Supervised learning is basically that kind of learning which we can used for labelled data or previously available data. (i.e. Historical data)  And, It is further divided into two classes:- 1. Classification (Finds the relation between discrete input & output) 2. Regression (Finds the relation between two or more independent variables)
  • 5.
     Decision Trees This technique is most powerful technique which used for classification.  It is a flowchart-like tree structure where each internal node denotes the feature, branches denote the rules, and the leaf nodes denote the result of the algorithm.
  • 6.
  • 7.
     Decision Treecan be Interpreted in following ways: -  Visualization  Tree based  Program based
  • 8.
  • 9.
  • 10.
  • 11.
     Preliminary Conceptsused in Decision Tree are : -  Entropy  Gini Impurity  Information Gain
  • 12.
     Entropy  Entropyis the measure of the degree of randomness or uncertainty in the dataset. In the case of classifications, It measures the randomness based on the distribution of class labels in the dataset.  It lies from 0 to 1 as 0 we can say the data is homogeneous. When its goes to 1, we can say there are lots of uncertainties in a data. (Multiple classes).  It is formulated as: -
  • 13.
     Gini Impurity Gini Impurity is a score that evaluates how accurate a split is among the classified groups.  It is also lies from 0 to 0.5. 0 as observations belong to same class and 0.5 is random distribution of elements with classes. So less Gini impurity score is highly recommended.  It is formulated as: -
  • 14.
    Purpose of Entropyand Gini Impurity is same Only difference is Gini Impurity has max value 0.5 and Entropy has max value 1.0
  • 15.
     Information Gain Information gain measures the reduction in entropy or variance that results from splitting a dataset based on a specific property.  The higher the information gain, the more valuable the feature is in predicting the target variable.  It is formulated by,
  • 16.
     Hyperparameters Hyperparameters comingunder Decision tree are: -  Pure node  Maximum number of leaf nodes  Minimum Information gain  Maximum Depth  Minimum samples at leaf node  Minimum impurity to split
  • 17.
     Random Forest This technique is used for both classification and regression.  This technique is based on a concept of ensemble learning, which is a process of combining multiple classifiers to solve a complex problem and to improve the performance of the model.
  • 19.
     Ensemble models •Ensemble – Group of things Combine different ML models to create a more powerful model. Following Ensemble models are: - 1. Bagging 2. Boosting 3. Stacking 4. Cascading
  • 20.
     Bagging –Bootstrap Aggregation • Bootstrap – Sampling with replacement • Aggregation – Combining • Random forest is a Bagging based ensemble model • Random forest can be used for Regression / Classification
  • 21.
  • 22.
     Bagging (Incase of classification)
  • 23.
     Bagging (Incase of regression)
  • 24.
     Hyperparameters Hyperparameters comingunder Decision tree are: -  Number of base learners  Decision Tree hyperparameters
  • 25.
     Why Randomforest?  No Overfitting 1. Use of Multiple trees to reduce the use of overfitting 2. Training time is less  High Accuracy 1. Runs efficiently on large database 2. For large data, it produces highly accurate predictions.  Estimates Missing data 1. Random forest can maintain accuracy when a large proportion of data is missing
  • 26.
     Naïve BayesClassifier  It is a classification technique based on Bayes’ Theorem with an assumption of independence among predictors.  In simple terms, a Naive Bayes classifier assumes that the presence of a particular feature in a class is unrelated to the presence of any other feature.  Despite its simplicity, it often performs surprisingly well in practice and serves as a baseline model for many classification tasks.
  • 27.
     Bayesian Method Convert the data set into a frequency table  Create Likelihood table by finding the probabilities  Now, use Naive Bayesian equation to calculate the posterior probability for each class.  The class with the highest posterior probability is the outcome of prediction.
  • 29.
     Types ofNaive Bayes 1. Gaussian Naïve Bayes: - • Variables are continuous in nature. • It assumes that all the variables have a normal distribution 2. Multinominal Naïve Bayes: - • The features represent the frequency. • If you have a text document, and you extract all the unique words and create multiple features where each feature represents the count of the word in the document. 3. Bernoulli Naïve Bayes: - • This is used when features are binary.
  • 30.
    Pros:  It iseasy and fast to predict class of test data set. It also perform well in multi class prediction.  When assumption of independence holds, a Naive Bayes classifier performs better compared to other models like logistic regression and you need less training data.  It perform well in case of categorical input variables compared to numerical variable(s).
  • 31.
    Cons: If categorical variablehas a category (in test data set), which was not observed in training data set, then model will assign a 0 (zero) probability and will be unable to make a prediction. This is often known as “Zero Frequency”. To solve this, we can use the smoothing technique. One of the simplest smoothing techniques is called Laplace estimation. On the other side naive Bayes is also known as a bad estimator, so the probability outputs from predict_proba are not to be taken too seriously. Another limitation of Naive Bayes is the assumption of independent predictors. In real life, it is almost impossible that we get a set of predictors which are completely independent.
  • 32.
     Logistic Regression Definition of logistic regression: Logistic regression is a statistical technique used to model the probability of a binary outcome based on one or more predictor variables.  logistic (sigmoid) function: The logistic function, also known as the sigmoid function, maps any real-valued input to a value between 0 and 1. It is the core component of logistic regression and helps transform the output into probabilities.
  • 33.
     Logistic function Mathematicalrepresentation of the logistic function: The logistic function is represented as where Z is the linear combination of input features and their corresponding coefficients. Graphical representation of the logistic curve: A graphical representation of the logistic curve shows an S-shaped curve that asymptotically approaches 0 and 1 on the y-axis.
  • 35.
     Advantages andLimitations  Advantages of logistic regression: Logistic regression offers a simple and interpretable model, provides probabilities as outputs, and can handle both linear and nonlinear relationships.  Limitations of logistic regression: It assumes a linear relationship between features and log-odds, is vulnerable to overfitting with a large number of features and is limited to binary classification.
  • 38.
     K-Nearest Neighbor(KNN)  The K-Nearest Neighbors (KNN) algorithm is a supervised machine learning method employed to tackle classification and regression problems.  The KNN algorithm works by finding the K nearest neighbors to a given data point based on a distance metric, such as Euclidean distance. The class or value of the data point is then determined by the majority vote or average of the K neighbors. This approach allows the algorithm to adapt to different patterns and make predictions based on the local structure of the data.
  • 39.
     Distance MetricsUsed in KNN Algorithm We use following distance matrices for following algorithms: 1. Euclidean Distance: Euclidean distance can also be visualized as the length of the straight line that joins the two points which are into consideration. 2. Manhattan Distance: This metric is calculated by summing the absolute difference between the coordinates of the points in n-dimensions.
  • 40.
    3. Minkowski Distance Wecan say that the Euclidean, as well as the Manhattan distance, are special cases of the Minkowski distance. From the formula above we can say that when p = 2 then it is the same as the formula for the Euclidean distance and when p = 1 then we obtain the formula for the Manhattan distance.
  • 41.
     Steps forKNN algorithm 1. Selecting the optimal value of K 2. Calculating distance 3. Finding Nearest Neighbors 4. Voting for Classification or Taking Average for Regression
  • 42.
     Easy toimplement - as the complexity of the algorithm is not that high.  Adapts Easily – As per the working of the KNN algorithm it stores all the data in memory storage and hence whenever a new example or data point is added then the algorithm adjusts itself as per that new example and has its contribution to the future predictions as well.  Few Hyperparameters – The only parameters which are required in the training of a KNN algorithm are the value of k and the choice of the distance metric which we would like to choose from our evaluation metric.  Advantages of the KNN Algorithm
  • 43.
     Does notscale – As we have heard about this that the KNN algorithm is also considered a Lazy Algorithm. The main significance of this term is that this takes lots of computing power as well as data storage. This makes this algorithm both time-consuming and resource exhausting.  Curse of Dimensionality – There is a term known as the peaking phenomenon according to this the KNN algorithm is affected by the curse of dimensionality which implies the algorithm faces a hard time classifying the data points properly when the dimensionality is too high.  Prone to Overfitting – As the algorithm is affected due to the curse of dimensionality it is prone to the problem of overfitting as well. Hence generally feature selection as well as dimensionality reduction techniques are applied to deal with this problem.  Disadvantages of the KNN Algorithm
  • 44.
  • 45.
     Precision: - Precisionattempts to answer the following question: What proportion of positive identifications was actually correct? Precision is defined as follows: -  Recall: - Recall attempts to answer the following question: What proportion of actual positives was identified correctly? Mathematically, recall is defined as follows:
  • 46.
    The F1-score isa measure of a model’s performance that combines precision and recall. It is defined as the harmonic mean of precision and recall, where the best value is 1 and the worst value is 0.  F1-Score  Accuracy We can find accuracy by calculating the ratio of correct predictions to the total number of input Samples.
  • 47.
     Sensitivity andSpecificity Count up the total number of each label (TP, FP, TN, FN) over a large dataset. In ROC analysis, we use two statistics:  Sensitivity: - Can be thought of as the likelihood of spotting a positive case when presented with one.  Specificity: - Can be thought of as the likelihood of spotting a negative case when presented with one.
  • 48.
     ROC Curve AnROC curve (receiver operating characteristic curve) is a graph showing the performance of a classification model at all classification thresholds.
  • 49.
     Unsupervised learning:-  Unsupervised learning, unlike supervised learning, does not come with labels (no output vectors).  The objective of unsupervised learning is to analyze the structure of the data and extract useful information from it without any explicit indication of the expected result.  Unsupervised learning includes two sub-types: 1. Clustering 2. Dimensionality reduction.
  • 50.
     K-means clustering K-means is a partitioning (clustering) method that divides a dataset into ‘k’ distinct, non-overlapping subsets (clusters) based on similarity, aiming to minimize the variance within each cluster.  It is one of the most popular unsupervised learning algorithms that deals with well-known clustering problem.  Since it can handle large sized dataset, It is the most computationally efficient than any other clustering methods.
  • 51.
     How k-meansclustering works?  We are given a data set of items, with certain features, and values for these features (like a vector). The task is to categorize those items into groups. To achieve this, we will use the K-means algorithm, an unsupervised learning algorithm. ‘K’ in the name of the algorithm represents the number of groups/clusters we want to classify our items into.  The algorithm works as follows: First, we initialize the k points, called means or cluster centroids. We categorize each item to its closest mean, and we update the mean’s coordinates, which are the averages of the items categorized in that cluster so far. We repeat the process for a given number of iterations and at the end, we have our clusters.
  • 52.
     How dowe choose the right k value?  The Elbow Method is a technique used to determine the optimal number of clusters (k) in the k-means clustering algorithm.
  • 53.
     What doesElbow Method do? We calculate the Within-Cluster Sum of Squares (WCSS) for different values of k. WCSS measures the sum of square distances between the centroids and each data point. As k increases, WCSS decreases because clusters become more compact. However, beyond a certain point, the improvement becomes marginal.  Graphical Approach: We plot k against its corresponding WCSS value. The graph often resembles an “elbow.” The optimal k value is where the graph starts to look like a straight line.
  • 54.
    From the abovegraph, we can observe that at k=2 and k=3 elbow-like situation. So, we are considering K=3
  • 56.
     Principal ComponentAnalysis (PCA)  In modern data analysis, principal component analysis (PCA) is an essential tool as it provides a guide for extracting the most information from a dataset, compressing the data size by keeping only those important features without losing much information, and simplifying the description of a dataset.  It works on the condition that while the data in a higher dimensional space is mapped to data in a lower dimension space, the variance of the data in the lower dimensional space should be maximum.  The main goal of Principal Component Analysis (PCA) is to reduce the dimensionality of a dataset while preserving the most important patterns or relationships between the variables without any prior knowledge of the target variables.
  • 57.
    Before PCA &After PCA
  • 58.
     Principal ComponentAnalysis (PCA) is a technique for dimensionality reduction that identifies a set of orthogonal axes, called principal components, that capture the maximum variance in the data. The principal components are linear combinations of the original variables in the dataset and are ordered in decreasing order of importance. The total variance captured by all the principal components is equal to the total variance in the original dataset.  The first principal component captures the most variation in the data, but the second principal component captures the maximum variance that is orthogonal to the first principal component, and so on.  Principal Component Analysis can be used for a variety of purposes, including data visualization, feature selection, and data compression. In data visualization, PCA can be used to plot high-dimensional data in two or three dimensions, making it easier to interpret.
  • 59.
  • 60.
     In featureselection, PCA can be used to identify the most important variables in a dataset. In data compression, PCA can be used to reduce the size of a dataset without losing important information.  In Principal Component Analysis, it is assumed that the information is carried in the variance of the features, that is, the higher the variation in a feature, the more information that features carries. Overall, PCA is a powerful tool for data analysis and can help to simplify complex datasets, making them easier to understand and work with.
  • 61.
    Neural Networks  NeuralNetworks are computational models that mimic the complex functions of the human brain. The neural networks consist of interconnected nodes or neurons that process and learn from data, enabling tasks such as pattern recognition and decision making in machine learning.  It is composed of an input layer, one or more hidden layers, and an output layer made up of layers of artificial neurons that are coupled.
  • 62.
     Types ofNeural Networks: - • Feedforward Networks • Multilayer Perceptron (MLP) • Convolutional Neural Network (CNN) • Recurrent Neural Network (RNN) • Long Short-Term Memory (LSTM)
  • 63.
     Feedforward Networks A feedforward neural network is a type of artificial neural network in which nodes’ connections do not form a loop.  Often referred to as a multi-layered network of neurons, feedforward neural networks are so named because all information flows in a forward manner only.  The data enters the input nodes, travels through the hidden layers, and eventually exits the output nodes. The network is devoid of links that would allow the information exiting the output node to be sent back into the network.  The purpose of feedforward neural networks is to approximate functions.
  • 64.
  • 65.
     Components ofFeedforward neural network  Layer of Input  Hidden Layer  Output layer  Neurons Weights
  • 66.
     Multilayer Perceptron(MLP)  A Multilayer Perceptron (MLP) is a type of feedforward artificial neural network. It consists of fully connected neurons with a nonlinear kind of activation function, organized in at least three layers  Every node in the multi-layer perception uses a sigmoid activation function. The sigmoid activation function takes real values as input and converts them to numbers between 0 and 1 using the sigmoid formula.
  • 67.
  • 68.
     Multilayer Perceptronfalls under the category of feedforward algorithms, because inputs are combined with the initial weights in a weighted sum and subjected to the activation function, just like in the Perceptron. But the difference is that each linear combination is propagated to the next layer.  Multi-layer perceptrons (MLPs) have a wide range of applications in various fields like, in Image classification, Speech recognition, Natural language Processing, Finance, Computer vision, Machine translation etc.
  • 69.
     Convolutional NeuralNetwork (CNN)  CNNs are special types of neural networks inspired by human visual cortex used in computer vision.  It is automatic feed-forward neural network in which information transfers exclusively in the forward direction.  A typical CNN architecture consists three types of layers: convolution, pooling, and classification.  Here, Even-numbered layers are used for convolutions, while odd numbered layers used for pooling operations.
  • 70.
     The Convolutionallayer applies filters to the input image to extract features, the Pooling layer downsamples the image to reduce computation, and the fully connected layer makes the final prediction. The network learns the optimal filters through backpropagation and gradient descent.
  • 71.
    Advantages of ConvolutionalNeural Networks (CNNs):  Good at detecting patterns and features in images, videos, and audio signals.  Robust to translation, rotation, and scaling invariance.  End-to-end training, no need for manual feature extraction.  Can handle large amounts of data and achieve high accuracy. Disadvantages of Convolutional Neural Networks (CNNs):  Computationally expensive to train and require a lot of memory.  Can be prone to overfitting if not enough data or proper regularization is used.  Requires large amounts of labeled data.  Interpretability is limited, it’s hard to understand what the network has learned.
  • 72.
     Recurrent NeuralNetwork (RNN)  Recurrent Neural Network (RNN) is a type of Neural Network where the output from the previous step is fed as input to the current step.  In traditional neural networks, all the inputs and outputs are independent of each other. Still, in cases when it is required to predict the next word of a sentence, the previous words are required and hence there is a need to remember the previous words. Thus RNN came into existence, which solved this issue with the help of a Hidden Layer.  The main and most important feature of RNN is its Hidden state, which remembers some information about a sequence. The state is also referred to as Memory State since it remembers the previous input to the network. It uses the same parameters for each input as it performs the same task on all the inputs or hidden layers to produce the output. This reduces the complexity of parameters, unlike other neural networks.
  • 73.
  • 74.
     Types OfRNN There are four types of RNNs based on the number of inputs and outputs in the network.  One to One  One to Many  Many to One  Many to Many
  • 75.
     How doesRNN work? The Recurrent Neural Network consists of multiple fixed activation function units, one for each time step. Each unit has an internal state which is called the hidden state of the unit. This hidden state signifies the past knowledge that the network currently holds at a given time step. This hidden state is updated at every time step to signify the change in the knowledge of the network about the past.  Applications of Recurrent Neural Network • Language Modelling and Generating Text • Speech Recognition • Machine Translation • Image Recognition, Face detection • Time series Forecasting
  • 76.
     Long Short-TermMemory (LSTM)  LSTM networks are an extension of recurrent neural networks (RNN’s) mainly introduced to handle situations where RNNs fail.  It fails to store information for a longer period of time. At times, a reference to certain information stored quite a long time ago is required to predict the current output. But RNNs are absolutely incapable of handling such “long-term dependencies”.  There is no finer control over which part of the context needs to be carried forward and how much of the past needs to be ‘forgotten’.  Other issues with RNNs are exploding and vanishing gradients which occur during the training process of a network through backtracking.
  • 77.
  • 78.
     Figure-A representswhat a basic LSTM network looks like. Only one layer of LSTM between an input and output layer has been shown here.  Figure B represents Deep LSTM which includes a number of LSTM layers in between the input and output. The advantage is that the input values fed to the network not only go through several LSTM layers but also propagate through time within one LSTM cell. Hence, parameters are well distributed within multiple layers. This results in a thorough process of inputs in each time step.  Figure C represents LSTM with the Recurrent Projection layer where the recurrent connections are taken from the projection layer to the LSTM layer input. This architecture was designed to reduce the high learning computational complexity (O(N)) for each time step) of the standard LSTM RNN.
  • 79.
     Figure Drepresents Deep LSTM with a Recurrent Projection Layer consisting of multiple LSTM layers where each layer has its own projection layer. The increased depth is quite useful in the case where the memory size is too large. Having increased depth prevents overfitting in models as the inputs to the network need to go through many nonlinear functions.
  • 80.
     Applications ofLSTM Networks Language Modelling Image processing Speech and Handwriting Recognition Music generation Language Translation
  • 81.
     Drawback ofLSTM Networks Problem of vanishing gradients Sometimes Inefficient Different random weight initialization More expectations required
  • 82.
    Time Series  Seriesof observations (data-points) recorded over time-intervals (usually equal time- intervals)  Note that in practice, data points are always noted at discrete times → time-series is NOT continuous
  • 83.
    Components of time-series Trend:-Long term behaviour Seasonal:- variations short term periodic tendency Cyclic variations:- repeating oscillatory tendency Random, irregular, Noise, residual:-unpredictable instances
  • 84.
     Auto Regressivemodels (AR Model) Self Relation between two variables predicting value of one variable to another variable “So, Basically AR model is Predict future values from past self values”
  • 85.
    The equation ofAR model is given by :- Where, β is Coefficients, Xt is targeted/Predicted value Xt-k is Pervious values εt is error at time t
  • 86.
     What isthe effect of AR model?  Autocorrelation function (ACF):-  Gives Direct as well as Indirect effect.  We estimate ACF by using Person’s Correlation Coefficient.  The equation of ACF for AR model is
  • 87.
     Partial-Autocorrelation function(PACF):-  Gives Direct effect only.  We estimate PACF by using Yule-walker’s Equation.  The equation of PACF for AR model is
  • 88.
     Applications ofAuto-Regressive model Benefits of Autoregressive Models:  Simplicity: AR models are relatively simple to understand and implement. They rely on past values of the time series to predict future values, making them conceptually straightforward.  Interpretability: The coefficients in an AR model have clear interpretations. They represent the strength and direction of the relationship between past and future values, making it easier to derive insights from the model.  Useful for Stationary Data: AR models work well with stationary time series data. Stationary data have stable statistical properties over time, which is an assumption that AR models are built upon.  Efficiency: AR models can be computationally efficient, especially for short time series or when you have a reasonable amount of data.
  • 89.
    Drawbacks of AutoregressiveModels:  Stationarity Assumption: AR models assume that the time series is stationary, meaning that its statistical properties do not change over time. In practice, many real-world time series are non-stationary, requiring preprocessing steps like differencing.  Limited to Short-Term Dependencies: AR models are not well-suited for capturing long-term dependencies in data. They are primarily designed for modeling short-term temporal patterns.  Lag Selection: Choosing the appropriate lag order (p) in an AR model can be challenging. Selecting too few lags may lead to underfitting, while selecting too many may lead to overfitting. Techniques like ACF and PACF plots are used to determine the lag order.  Sensitivity to Noise: AR models can be sensitive to random noise in the data. This sensitivity can lead to overfitting, especially when dealing with noisy or irregular time series.
  • 90.

Editor's Notes

  • #5 https://media.geeksforgeeks.org/wp-content/uploads/20230424141727/Decision-Tree.webp
  • #6 Fundamentals of Machine Learning by Dr. Nileshchandra Pikle
  • #8 Fundamentals of Machine Learning by Dr. Nileshchandra Pikle
  • #9 Fundamentals of Machine Learning by Dr. Nileshchandra Pikle
  • #10 Fundamentals of Machine Learning by Dr. Nileshchandra Pikle
  • #14 Fundamentals of Machine Learning by Dr. Nileshchandra Pikle
  • #17 data:image/jpeg;base64,/9j/4AAQSkZJRgABAQAAAQABAAD/2wCEAAoHCBEREREQEBEREBARExERERcZERkRERERGBoaGRcZFxcaICwjHR0oHRwaMTckKC0vNDIyGSI4PUUxPSwxMi8BCwsLDw4PHRERHC8oICgzMjEvLzE0MjE4NDM1LzEyLy8xLzExPDoxOjExNy8vMS8xMTExMS8xMTEvMTEvMS8zO//AABEIAK4BIgMBIgACEQEDEQH/xAAbAAABBQEBAAAAAAAAAAAAAAAAAQIDBAUGB//EAEUQAAIBAwIDBAUKAwUHBQAAAAECAwAEERIhBRMxBiJBUWFxkqHSFBUWMlJTVJGTlCOB0TNCVaSxNENyc3Sz4SRigsHw/8QAGgEBAQADAQEAAAAAAAAAAAAAAAECAwQFBv/EAC0RAAICAQIEBAUFAQAAAAAAAAABAhEDEiEEMUFRBRNxoRVSYYGxMkKR4fAU/9oADAMBAAIRAxEAPwDzmeTBPrNRGb0e+luDufWagrXGKaPb4riZxm1F7D2eo2NFIaz0o4pcRN9RpFNK040lXSjRLJJ9RhT000x+n3VLQatI1uTIeV6fdRyvT7qlopSJZFyvT7qTk+n3VNRSiWQ8r0+6jlen3VNSUpCyHk+n3U9LVmOFDMR1AUsfdVqxtjNLHEDjWwBPkOpP5A16HbW6RIEjUKo8B4+knxPprt4Tg3nt3SRx8Vxaw0qts81+bpfu5P02/pS/N0v3cn6bf0r1O3hLlgCqhF1sWbSoXKr19bD86k+QTZK8p8guDtlcqCzYbodgenXG2a634bjXOZyrxKb5R/38HlHzdL93J+m39KPm6X7uT9Nv6V6obWUbmKQAAMe4fqnOD7j+R8qJbSRFLvG6KGCEkY7xGQN/RT4Zj+cfEZ/L/v4PKJLJ1GWV1HmUKj8zUfI9Pur1RhkYO4OxB3BFcR2j4esMo5YxHICwH2WB7wHo6fnXPxXAPDHWnaOjhuNWWWhqmYXI9Puo5Hp91TUVwUdxByPT7qOR6fdU9FKBByPT7qOR6fdU9FKBDyPT7qOR6fdU1LUoEPJ9Puo5Pp91T0UopDyvT7qOV6fdU1FKBDyvT7qRo8DOfdU1Nk6GlCyvRRRWINSc7n1moqfKdz6zTKQ5HfxTvIwppopDWRyMSiiihgwpKUDw8TsPMmr3zJefg7v9tL8NDEoUVf8AmS9/B3f7aX4aPmS9/B3f7aX4aAoUVf8AmS9/B3f7aX4aPmS9/B3f7aX4aAoUVf8AmS9/B3f7aX4aPmS8/B3f7aX4aoK9hc8maOXGdDAkea9Dj04Jr0O3nSVQ8bBkboR/9+R9FcJ8yXn4O7/bSfDUkXCb9M6La+TPXTBKufXha7OE4x4LTVpnFxXCLNTumj0C3nMZYhUYOuhgy6lK6lbp61FWG4pKRhirAlicr11BwQQDjo79MeHkMeefIuJ/c8R/Sn/pSGy4pg4h4hn/AJU/9K6n4jifOBzLw+a5TPRzxibc5UHGOh23LZAJx1JxtgbYxgVDc30koIfBBKsOvdI1dMk4HeO3+lch2gseIfKX5cF7o0Q6dEUunPKTVjSMfWzn05rO+RcT+54j+lP/AEovEMS3UCvgMjVOR2prie0l+s0oEZ1JGCobwZicsR6Nh+RpsvDeIuMPb37jyaGZh+RFQ/Ml7+Du/wBtJ8NaeK4/zoaIql1N/C8EsMtTdsz6K0PmS8/B3f7aX4aPmS9/B3f7aX4a887jPorQ+ZLz8Hd/tpfho+ZLz8Hd/tpfhoDPpav/ADJefg7v9tL8NL8yXn4O7/bS/DQGfS1f+ZLz8Hd/tpfhqrPBJGxSVHicYJV0KOAemVYA1CkWKXFKBS4oWhuKMU/FIRQUMNMk6GpDTJOhoQrUUUViDQc7n10ylJ3PrpKseR153c2FNNLSVTnYVLDayyBjHHJIEGXKozhR6cDaoq7jh3E1soWjjiB02VvfM2s6pJZWiRgc5AA5i4x4J4k5qqupz55TjG8at9ro5Hh181uzyR7SlGjjcHDQsxXLoftaA6g+GvI3AquZGO5ZifHvEk1d45EiznlqEWSK1m0gnSjTQRyuFJydIZ2xknbFUKhmLzG+03tGjmN9pvaNJRVKLzG+03tGjmN9pvaNNooB3Mb7Te0ascPlkE0TIJJXV0dUBYl9JDYGM+XlVWuo7IyLCFn5YeWW8hswSSOXHIpLFQNtWoL1zsCNs5oYZHJRbirfRGDfyyGaUuHidnd2QlgU1Etg5x5+VV+Y32m9o10XaidLhBc8sRutzNaHBJ1xxojAsDtqyzdMbEDfGa5ugg5OKclT6odzG+03tGjmN9pvaNNpaGwXmN9pvaNHMb7Te0aSigF5jfab2jRzG+03tGkrQ4JGjTguodY4rqbSSQrNDBJKgbBB06kXOCNs0AvFZp2EBkilhCQRQqW1DmhM98ZA656b1n8xvtN7RrueK8RW7iEUkQGbGS/VtZ1RyprVQMYBHdbOeuodMZrhau3Q1YHOUbyKn2uxeY32m9o0cxvtN7RpKKhuF5jfab2jS8xvtN7RptFAP1t9pvaNW2v3eAW8mqQJIrwszZMK4YOi530uShxkAFAfE1SFKxwCfIE1C0TpbSFDII5DGpwzhCUB9LYxTAK7a94ulpDeW8durRW04tFVpGIkSQTMzOeuTymzgj6+2MVy3FrZYbm5hTOiG4niTJydCSMq5PicAUdGrFLJJyUo0k9t7tdylikIqTFIRUN9ERFRSdDU7ConGxqmDRUop2mlrEuks53NFNU7mlqx5G3K7kwpfQASTsANyT5CgAkgAEkkAADJJPQAeJrUbgMqsUlls4CraZNV7CWjI+tqRHL5G/dAJ2wBmqaWzRv+xk1qizXk0MMJwGK65pA5GQgjCgM5wf7wXY5IqvJepMLwxK6RxcPt4I9ZBdkjuLVQzY2BPXAzjzPWrXbbjMs7xW/NkkgjhtZAXRUeWWSFHMjBehw/1cnG/rq/2D4PZ3FtetcS6GYcpxzAnLhBjkEm/wD706nbuY8TXN5ksWPVl3f09djPSpSqPuYnIjv2j5MnKuhFbwmKQHlycmJY9Ucig4JSMEo4Hjhm6VS4vwiW0dUl0HWupSpLKQNj1AOR6vGq0M7RSCSJyGjYlHAxnGwOD5jwPniuk7RQG6NvKtxDra2tmeKWaK3CM8YLPGZGCshfVkZ1KRjppNdW1HLJZPMVVprfvZytFWr2wlg0GQIVkDGN0lSeN9JwwDxsVyDjIzkZHmKq0NwUUVeg4TM6JJmBEkBKcy6hgZ1BKllWRwSuoMM4xlT5UIaXCux93dW/ypDCsWWxrkZWZVJDMAFOwIPp7p2pbG4hR7S1gd5839tcSSleXGWBCBY0PeK9495tJP2RV6XiEtnwtYIroO0s8kcmgq8UUfLDyRxyDOo5kTUynALEL4k5HZOCN7yISPoCHmJuBqdSGRcn07/yrVjWTU9VVe1dvqTNOEIat9lb/oc1xC6zWs7tBi7nnjlC82IMwCFZEXvBe6O8uojJ7pqbi3Y+7tYPlMpiaLKg6JCzKGOFYgqNiSOhz3hVTtPaxRXUqQtrU99tw2l2yWXP/wC61tJxCa84UYZboI0VwkcesokUqcvXHHJIcaTlHKsxwSuG8CGRZNS01V732+hcM4zjq33Vr+zkKK0J+ESojy5gdE0l+XdQzsisQoZljckLqKjOMZYedZ9bShRRVuy4fJPrMegLGFMjvLHBGuo4UF5GAycHAzk4PkaFJeEcIlu3ZItA0KGZmJVRnYDYE5O/h4VcMEdi0glfm3RiuIhHGDyoudE0RaSRgMkLISEQHwyw6VodnYDa/KZWuIda2tw0ccU0dwGZEJR5GRiqqrlMDOpifs6q5qWZpZC8rnVI4MjkZxk4LYHkPAeVNqNMVkeR21prbvZuPeJCbUyq7Ry8M5EmggOqyPMpZdWxI8jjPmOtSWPYye6jaa0mhmhGQpbXFIzDqhQqQrjb+8V3GCa0u33CLSCCzaCbW4VYVHMV+ZAA78zbyY9Rt3qz+xnGZYmltuc8UEkN1LlUV2iljiZ9aA9TpQ7ZGcDod65vMlkx6sWz+vqdWlRdS9jl8egg+XQj10lay8BlZgkUlnMWIWPRew6pSdl0ozh8nbCkA74xmsogjYggjYgjBB8QRXSYCUtFTWlu80iRRKXkkYKijAyT6TsB5k7AAk0AW1u0siRRjLuwVR03Pp8q2bvs98mwb2dEjYHSsStLNLjGoIGCquMjLMRjIwG6VFZ8MZJEke5tolRg+qO6hnmyOgjijclnJ2A6b7kDJp/ay9eW8nQsTFBNPDCNu6iuVzkdSdI3ptRql5jyR0taad97DiF18ohvp9PL517ayac6tGqO9OM4GceeBV1OGDidxLJaSKkszyzyxS6lMOttTFJEUrIgZsdFbcbeNaHCeDWb8Hmlkm0ykmZu+By5ohKsSaepyHO3U69vCuW4bxOa0kE8DaJFUjoGDKeqkHqDgfkK5vN1qShzW2/Kzq06Wm+TJOLcNltZmt5godQDlTqRlIyCpwDj1gdDVM103aqzM17cSJcQuOY66ZbiK2kiCnCrpdlDJp0lWXqDvhtQrn7y0khfRIACVV1IdZEdGGVZHUlWU+YJ3BHUGtkdWlaufWiNq9iswqJxUzVGRWZjW5BpoqTTRWuzZpI0O5p9RR9TUtbUapvc7rsJZ8OMEtzcTLHcxM5Rmk0mBdI0SRoTgsCSQSDhgMbiuDHT+VWbGze4mjhj08yRtK6jpUHBO58BgGp5+D3CZYRmWMIJeZH/ABoTGSwDB19KuMHBGhs4wa1wg4ylJtu+j6ehJTTSVcjrOP2nD34XFdCVXveXbpq5n8SV1CoyyRg4yq53xnuDJPjwhAPhV9+CXSgs1vMCNepeUwdQgRmLDGw76/zOKWTgt0raeRM/feIMsTsjOmosFON8aW9hvI4YoOCabb3vf8CUlLkqLnZC0tZrtIrxwsRVyAX0LJJtpQuCCM79CMkAeNSdtuQLwx2zh4Yo4okCkMkWgYKKw6gdSdzqZsknNZsfBrp+aFt5mMUiwyqIyXSRgxClOucI3htgZ6jIvCLnKqLeclkLoBEx1IMZI26d5fbX7Qy0PzNVuqqunqTUqqvubvZSG1mt54bqRdIkWVUaTlhMIVLocg5IJB8DpTPQVzEwUMwQlkDMFJ2LLnYn1ipfkMht/lOFMXNEIOtS3MKs31euMKd//OK9bbNEcbjOUnJtOtui9Dd7GWdrPdiO9YLEUYqC5jWSUFdKlgQRsWOxGdIo7aGH5dIIJBJEixRpggpGEQLoQjbSuP8AXqcmsFjsc9K1rns5dRtoWMTEPJE3JbnaJI8a0cLurDI6jfO2a1eW/M1W6qq6ep0alp019zo+xNpw+e0njvpUHLlMqK0oiMKmNVZ0YEE6sAEZI/hpt58Q4XLBTqXJCkjBZc7EjwyPCr0fBbl1DiCXBUOmY2BkUkLlNtwMgk+RzSHg9zp1CCVlC6mKxs3L77oQ+2x1I23o9eEMbjKUm276dvQjkmkq5FONV1KGJVCyhiBkque8QPHAztXa9trOwgtYIrGRDzJRK6rIJTKoRlV3Yknu5IAyB332rlvme51tGbeZZEiedkZCjiJfrPhsZA3/AJgjrtQeDXSnBtZwS/LxynB14zg7dcUljcpRkm1XTv6lUkk1RodizD8ujWeQRwuksb5IVJA6FdDk7aWz+YHjik7Y2drDdtHZMGiCIWAcyLHLvqUOSSdgp3JwWIqjZcJmmlaBdEcynTokcRSO52CKrblj+XmRVBegx08KaH5mq3VVXT1LqWmq+4+FVLqHJVCyhyNyqkjUR6hmun7VxWsUEENrIunmPKyLJzA+pQokc5JzgADwGpsdTXKitKHgk7wrOgjYMryLGHHOaNGKO4j6kAg9Mnx6Vts55Y3KcZamkr26P1L/AGJEDXnLuXCQzRSxOGIVJdQGEZj0ydwRg6lXBBxUPa20tobt47Nw8QVCcNrVH31KHJJIG3icEkeFVk4HdM2j5PMp1RoxaJgqtIQE1HHiTTF4PcnAEEmpjEFTltrYSLI6kLjppick+GPXjVofmardVVdPU6dS01X3KIUeAru+B2fD04XLdGVUveVcIW5n8SJ2DKqJGTjLLgZxkhzgjw46Xh1wih3gmRGflKWiZQZcsNAyPrZVhjrlWHganl4Fdo7I1rPqQyA4iLD+G2h8MMhgG2yCRmmWDmkk2qd7fgRklzVmf7vV1Hqrue3lnw8QxXFvKslzK6lysmszqVOuR1BwrFgCSAMljnc7ch83zB4kdDEZscsyDlIwON9TeG4/MeYycQsnt5XglCiRNOrScghlDqc/8LDruOhwaTg5SjJNqunf1LGSSaoq11lhDaJw6WZZQtzJDLGx1/xAzdY1U9AwGkkDJVm3wTXK1d4dwuS4DmMxDQ0UY1yrEXkk18tELbFjobqQNq2pmjPic0qbVNPbr9CpCcMp1FMMp1D6yYI7w9I6/wAq7Tt7ZcPVI57WRGuJnLSaJQ4lRlYtIVBwp1Y+qAO8dvLlk4RdEqBbTkuGKjkvkhcaj08Mj8x5invwW5XOYXC6WcNobQ+I+YQpxudIP8wfI1qnBykpJtV07+pvUkk1RQwM5wM++uo7D8Psp5ZVvWXCohiRpTErnJ1kkEE4AXbP949cViycIuk1l7adREuuQmFwEXBOWONhgH2T5Gj5nuiqMLeZkkSORCsZdSkhIQ5XOMlWwDvtmmWLlFxTavqiRkk7Je0EqPd3Lxyc6NpXKPthl6DGABpHQY2wBiun4PZ8Pl4UWuphzoflBTMmmSAscqqID31YgMFIPed8YJNcdPYzxI0kkMscasUZmjZVD+RJHWrPEuC3FuC0yBQDCoOoMG5qO6FSOoxG4J8CpHWsJ4nKKim1VbrmzJTSd0Z+abSmkFbmRc0JiinYorWdFFOPqalqGLqamrajib3LFhePbzRzR6dcbal1DUucEbjxGCa007U3aKY43SKPAEaxpykhA1f2YUjrqbOrVkkk771iUUIbKdpJ1LlFgQPziyrHpQtKEEhK6sHPLXY5X0YwA9uP3kqT/wBmRy5eewQB+TKzoVJz9XXcNgDfLjqF2wya1OE3NskV0s63BeaJY4+WyBTiSKTDFhld4x3hq2J223AROOThp3/hlrkqZMpnDCOSMFN9jokceP1qtx9rr1dOJR3YxH/fGQujSxwwww5abrjODnOTWAD6R/5paAtSX8zxmKSRpFLxSEsS75jR0Qaj4BXbb1VWpKWqBGGQR51vjtZdiRpFMUesyNIqR8tJZJGVmd9JDFtSLvq2xjoWBwaKgNeLtJcKQ45Zk0RxvJpIllWPSE1urAkgIoBGD/M5qaLtHeSMsaiFnd3WIclQUeZn1iM9F181lOfDHQjNYVXODXEMVxDLOJGijkSQ8srrypDKQG2IyBkZGfMUBLJxacSl30FxHcW7AplSkryPICAeuuV8EHwHlvcj7V3q/VkA/iSSdCB/EZ2kQgHBQs7npkZ2IwMZPEJImmlaHWI2YsocrzN9yW07DJztvgEDJxkwUKbMPaW5WSSQlJTIrJiQNKIkf66x5bKBhgHB3AwaxgKKKAK1rTj9xDHHHEY0MWoRyCJTMis/MZQ7ZwC3o3G3QkVk0VSm3L2muXdXYRO0cizQ6kL8h1KkmMsxKhiq5GSNtsUn0luSAp5TJoSJlMeoSRqsqBXJOojEr7AgDbGMVjUhNQGtxbiN1Kum4xicJMMKoLKHnZWGOg1Sy7egeVT/AEpug7uGRWdSrBVKDJkebUMNkNzJHPXHexjFQcbu7aVbcW6zqYoRE/MKYwGZlC6PrHvHvd3O3dFZVCmxcdo7qUxmSTUE05U5MUpGneSMnST3V6AbjOx3qlxC9kuJWml063CA4GFARVRQB6FUdSScb1UpaAdV3hvFJbYOI1ibW8Un8SIS6JY9eh0B2DDW3UHrVDNGaoNlu0tyyrHIyTIIxEyyKZElQaCvMUtgleWmCMHu75ycuHai7AUB0UpuraMsp5bRAjJIB0seg3OM5wAMTNJmoNjbm43dSRsdMSwoWiwkYREM0c6kAZz3lec+OD5bCmW/H7iNYVQovJ06To7xVRMoVjncaZnG2DgjxGajhubYWc0Tif5Q8sUi4Kcn+GsqgkkahtIcrg5wN1ztmBhQG1cdpryROWZmUaiVKu6uqkk6A2rdck9cn042qK+43czxmGVw0ZlE4GnGmQIY+6fBcE7dMnPic5maKAWlFJSr1qPkbI80OxRS4orWdmkz4epqaoYepqatqPMYUUUVSGpwPjT2ZmaNQxlj5Ry2NIznI2IJ9e3nnpV09oICTnhttgyLIANI0gBe4O59Uld857pI2ySc7gFukt3bxSLqjeQK4yRkYO2RvWah2HqFQpuXfaFpGL8lEdoJrdyGOWjk5hwcYBALg7gklBuBtWLRRQBRRRVAtFJS0Ba4ZetbTRzoAWj16QTgZZGXO3lqz/Kt6PtCtxOmqwtnlmk0nVjDvJIpUuShbA6dd8/yrl6ucE/2q1/6m3/7i1Aadxx1QIYTbIEtWjC6ZmSTMZfUwkRVIdiSdXTJJwSRjJ4jdm4nlnIIMrl8FtZHo1ADPrx/WpLyyn5sv8Cb+0k/3L/aPoqL5DP9xN+i/wDShSvRVj5FP9zP+i/9KhkjZDpdWRuuGUqceo1QNooA8BuT0qx8in+5n/Rf+lCleukftVrSNJrWGYRJBHGWw2lYwAca0b6wAyOndXbY6sP5FP8Acz/ov/Sj5FP9zP8Aov8A0qA6C440IUXFlBGt1DLIQr/Wjk5kKg90EadLHqd2JyM7Y3EOJGaO3i0lVtw4XLmTAfT3VyMhBp2BLEZxnAAE3G42VLFWUqws1yCCpH8e46g1lUIFLSUVS2LSUUUFhW9w3tKYIEtnt4p4VdpMOdmZtXgVI2LA9P7orBoqEN9O0Ma4KWUKSiGSEOCoGXVVZyoTc4XAAI2dwc6iaSXtIxiuIBBEiXDzyHBYsjymIkgnxxHggYU6hsNIrLjgU200pzrSa2iXfbTIlwzZHnmNPf51VFAOzTqYKcKGaFp6dRTBT06ijM4fqRLiilorUelpMqHqanqGHqamrajxQoooqg1Oy/8At1r/AM0f6GshOg9QrX7MMBfWpJCjmruSFAzkbk7CkXs5eAAcleg/38Xx0BmUVq/Ry8+6H60Px0fRy8+6H60Px1CmVRWr9HLz7ofrQ/HR9HLz7ofrQ/HQGVS1qfRy8+6H60Px0v0dvPuh+tD8dAZVLGrFlCBi5I0gZ1avDGPGtT6O3n3Q/Wh+OkuLGa2hZpFCPKwiGHRyI8FnwUY4yQB6gfOs4JNmMpUth3Ouhs18UPiDeuSD6dJI99HPuf8AEP8AOyVkgV1PCOya3EEVx8o0KyXDzLpBZCrOkITfva2jfPkFNXVH5fclPuZnPuv8Q/zstP4n/HW1zcxSSRQGKVnlZmL86aQd4gkjQ6f6eFWpOyMyhi0sPd0szBmZY1C3BlDKE160MBGkA9fXpfF2Qk/iLLPFG6LI2FEkuBGZ1OoBB9ZoGxjJA6jO1NUexdL7+xm8LhWK4gleeDRFNDI2JcnSjqxwMbnAqSW5uCzEcQ2LMR/6yQbE7bUN2dnEsVuGieaaOSaFVYtzY11FCh04/iKrFc4yAM4yKtN2VdUkdriHSkbyZUSOrgRySppbQAwdInOQTp2zucVNUOw0y7lPn3P+If52Sjn3P+If52StC57Hyo8yiaArEz6/7RnSMc3QxVU7xPJfZM426Dpz99atBNNA5BeGR4nIyVLIxVsZAOMjxAq6o9hUu5JepNkSSu02ruiTm80Njw15P5GqtXOF96QQ/wB2YGNh6cHS3rVsEH11OOzt59yv68Px0klSaEW7aZmUVqfRy8+6H60Px0fRy8+6H60Px1rMjLorU+jl590P1ofjo+jl590P1ofjoDKorV+jl590P1ofjo+jl590P1ofjoCGH/Y7j/qrH/tXlURWzPw+WC0lEqqhkubQoOZG5YJHdajhGOw1L+YrGoEKKcKaKUVTNDqenUUwU5OoqM2Y/wBSJ6KTNFaz1DMh6mpqhh6mpq2I8NhRRRVAGm6B5D8qdRQDdA8h+VGgeQ/KnUUKN0DyH5UaB5D8qdS0A3QPIflRoHkPyp1FCjdA8h+VWbSRAGjfIjk0nIGTG650sB49SCPEGoKKsZU7I42qLfyLylgYeB5yp7nwR+VOW3kGAJogB0AukAGc5x3tup/M+dUqKtx7e5NMu/sawlutIX5VHhWSQH5VHzAyKypiTVrwAzYXOBqbbc1AkMykFbiNSNwRdqCDknqG82Y+tj5mqNKsbMQqKzOxCooBZmY7AADcknwq3Ht7in39i/MkzuJGmh1qqKpE8cehUAVAoQgKAAOmKRYplChZ41CatAF2oCatn0gN3c75x1zvUvaO05VwwEfKjZIWj7mhGHLTUV2we/qzjxzWVS49vcVLv7GpbtcRyCVLiHmDcFp45RnDDJWQkE4ZtyNtR86hmtpJHeR5YXeRmd2NxHlnYksTv4kmqNFLj29yaZdy4umEEh1kmIZV0HKRhhgsW8WwSABkDOfRVDQPIflTzSVJSv0CVeo3QPIflSaR5D8qeaSsTIbpHkPyo0jyH5U6ihBNA8h+VGgeQ/KlooUAAOgA/lTqSigFpRSUUMkyShTTM0uajNkHuifVSVFqpaxO7WVIepqaoYupqaskeWwoooqgKKKKAKKKWgCiiihQooooUKKWigNLhXB5LjLAhIwcFiM5Pko8a2Poqn3z+yK1OCgC2h09NCk/8R3b35q43Q17+DgMXlpyVtng8Rx2XzGoukjn/oqn3z+yKB2WT76Qf/EZFdvci0ZpdHdWNjoCnQZFAkPdZi+QSI98eI2pJLW2VJCsgkYc0J31IYZfQQAAQe6vXY69qiw8P8jJ/wBPEfMjjrrgDTOZJ7qeaQgAu55jkDYDUxJwKi+isf3z+yK7KK2tyhJlYOFQ41puzJqbAYDIDd3AORjPoqZLK2EikyqYw41AyISUzFgEqBnZpMlcgaBv5nh4Zfs/IXE8Q/3fg4b6Kp98/siqnEOzTxqXifmBRkrpw+PHGOvq2rtLlUDLo+qUiJ74c6yilskDY6s7ejyxUFbHwOCcbUaMY8dmjLd2eZ0lWL8KJpgv1RJIB6tRqvXz0lTaPfTtWFJS0lQBRRRQgUUUUAUUUUAtLmm0UKOzQTtSUjnY1DJPcNVFQ5orE3axYupqaq2ojpTtZ86qZoonoqDWfOjWfOrYonoqDWfOl1nzqWWialqDUfOlDHzpY0smoqLUfOl1HzpZVFklFR6zS6zSzJQZJRTQacDTUZrEzX4PxpoBy2XXFkkAHDIT1x5j0VsfSe3+xP7KfHXIilxXXj4/LjjpT2OafheLJLU1udd9J7f7M3sp8dH0nt/szeynx1yWKCK2fE8v0J8Fgdb9J7f7E3sp8dH0nt/szeynx1yJFNNPieX6fwYvwbGuh1/0ot/sTeynx1Uv+02pSsKMpO2tsAr/AMIBO/prmTTSxrGXiOaSqyLwvHF3Q+g1GWPnTSx864rOhwZLSVCXPnSFz51bMWieiq3OPn/pRzT5+4UsxLNFVeafP3Cjmnz9wpZC1RVXmnz9wo5p8/cKWC1RVXmnz9wo5p8/cKWC1TX6Gq/OPn7hSlz50soZoptFYmVn/9k=