KEMBAR78
Classification in Machine Learning | PDF | Statistical Classification | Machine Learning
0% found this document useful (0 votes)
6 views4 pages

Classification in Machine Learning

Classification is the process of recognizing and grouping objects into predefined categories using machine learning algorithms. It includes two types of learners: lazy learners, which store training data before making predictions, and eager learners, which build models from training data before testing. There are various classification tasks such as binary, multi-class, multi-label, and imbalanced classification, each with specific algorithms suitable for their respective tasks.

Uploaded by

ghugekrish824
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
6 views4 pages

Classification in Machine Learning

Classification is the process of recognizing and grouping objects into predefined categories using machine learning algorithms. It includes two types of learners: lazy learners, which store training data before making predictions, and eager learners, which build models from training data before testing. There are various classification tasks such as binary, multi-class, multi-label, and imbalanced classification, each with specific algorithms suitable for their respective tasks.

Uploaded by

ghugekrish824
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 4

What is Classification?

Classification is defined as the process of recognition, understanding, and grouping of objects and
ideas into preset categories a.k.a “sub-populations.” With the help of these pre-categorized training
datasets, classification in machine learning programs leverage a wide range of algorithms to classify
future datasets into respective and relevant categories.

Classification algorithms used in machine learning utilize input training data for the purpose of
predicting the likelihood or probability that the data that follows will fall into one of the
predetermined categories. One of the most common applications of classification is for filtering
emails into “spam” or “non-spam”, as used by today’s top email service providers.
In short, classification is a form of “pattern recognition,”. Here, classification algorithms applied to
the training data find the same pattern (similar number sequences, words or sentiments).

Learners in Classification Problems


There are two types of learners.
Lazy Learners
It first stores the training dataset before waiting for the test dataset to arrive. When using a lazy
learner, the classification is carried out using the training dataset's most appropriate data. Less time
is spent on training, but more time is spent on predictions. Some of the examples are case-based
reasoning and the KNN algorithm.
Eager Learners
Before obtaining a test dataset, eager learners build a classification model using a training dataset.
They spend more time studying and less time predicting. Some of the examples are ANN, naive
Bayes, and Decision trees.
There are four different types of Classification Tasks in Machine Learning and they are following -
 Binary Classification
 Multi-Class Classification
 Multi-Label Classification
 Imbalanced Classification

Binary Classification
Those classification jobs with only two class labels are referred to as binary classification.
Examples comprise -
 Prediction of conversion (buy or not).
 Churn forecast (churn or not).
 Detection of spam email (spam or not).
Binary classification problems often require two classes, one representing the normal state and the
other representing the aberrant state.
For instance, the normal condition is "not spam," while the abnormal state is "spam." Another
illustration is when a task involving a medical test has a normal condition of "cancer not identified"
and an abnormal state of "cancer detected."
Class label 0 is given to the class in the normal state, whereas class label 1 is given to the class in the
abnormal condition.
A model that forecasts a Bernoulli probability distribution for each case is frequently used to represent
a binary classification task.
The discrete probability distribution known as the Bernoulli distribution deals with the situation
where an event has a binary result of either 0 or 1. In terms of classification, this indicates that the
model forecasts the likelihood that an example would fall within class 1, or the abnormal state.
The following are well-known binary classification algorithms:
 Logistic Regression
 Support Vector Machines
 Simple Bayes
 Decision Trees
Some algorithms, such as Support Vector Machines and Logistic Regression, were created expressly
for binary classification and do not by default support more than two classes.
Multi-Class Classification
Multi-class labels are used in classification tasks referred to as multi-class classification.
Examples comprise -
 Categorization of faces.
 Classifying plant species.
 Character recognition using optical.
The multi-class classification does not have the idea of normal and abnormal outcomes, in contrast
to binary classification. Instead, instances are grouped into one of several well-known classes.
In some cases, the number of class labels could be rather high. In a facial recognition system, for
instance, a model might predict that a shot belongs to one of thousands or tens of thousands of faces.
Text translation models and other problems involving word prediction could be categorized as a
particular case of multi-class classification. Each word in the sequence of words to be predicted
requires a multi-class classification, where the vocabulary size determines the number of possible
classes that may be predicted and may range from tens of thousands to hundreds of thousands of
words.
Multiclass classification tasks are frequently modeled using a model that forecasts a Multinoulli
probability distribution for each example.
An event that has a categorical outcome, such as K in 1, 2, 3,..., K, is covered by the Multinoulli
distribution, which is a discrete probability distribution. In terms of classification, this implies that
the model forecasts the likelihood that a given example will belong to a certain class label.
For multi-class classification, many binary classification techniques are applicable.
The following well-known algorithms can be used for multi-class classification:
 Progressive Boosting
 Choice trees
 Nearest K Neighbors
 Rough Forest
 Simple Bayes
Multi-class problems can be solved using algorithms created for binary classification.
In order to do this, a method is known as "one-vs-rest" or "one model for each pair of classes" is used,
which includes fitting multiple binary classification models with each class versus all other classes
(called one-vs-one).
 One-vs-One: For each pair of classes, fit a single binary classification model.
The following binary classification algorithms can apply these multi-class classification techniques:
 One-vs-Rest: Fit a single binary classification model for each class versus all other classes.
The following binary classification algorithms can apply these multi-class classification techniques:
 Support vector Machine
 Logistic Regression

You might also like