Q1 What is Business Intelligence?
• Business Intelligence (BI) involves analyzing data and using mathematical
models to generate insights for better decision making.
• BI helps organizations make effective and timely decisions by providing
tools and methodologies.
• It supports decision-making at all levels, from individuals to teams, in
both public and private sectors.
• BI enhances the performance and competitiveness of organizations by
enabling informed decisions.
• Modern decision-making often requires analytical and rigorous methods,
as intuitive approaches alone may not suffice.
• BI leverages available data to extract valuable information and
knowledge.
• It connects with fields like data science, statistics, and business
management to solve complex problems.
• BI provides easy-to-use tools for knowledge workers, enabling them to
analyze data effectively without requiring deep technical expertise.
Q2. Business intelligence Architecture
1. Data Sources:
• Data is collected from various sources, including operational systems,
emails, or external providers.
• These sources can have different formats and structures.
2. Data Warehouses and Data Marts:
• Data is processed using tools called ETL (Extract, Transform, Load).
• The cleaned and organized data is stored in databases to support
analysis.
3. Business Intelligence Methodologies:
• Data is analyzed using advanced tools and methods to assist decision-
makers.
4. Data Exploration:
• Basic tools like queries, reports, and statistics help decision-makers
analyze data.
• These methods are called “passive” because users need to define what
to search for.
5. Data Mining:
• Advanced “active” methods automatically extract patterns and insights
using techniques like machine learning and pattern recognition.
6. Optimization:
• Models at this stage help find the best solution from a range of possible
options.
7. Decisions:
• The final step involves choosing and implementing the best decision
based on all the analysis
Q3. Types of Decision
• Decisions in an organization can be classified into three main categories:
structured decisions, unstructured decisions, and semi-structured
decisions.
• These classifications are based on the nature of the decision-making
process and the availability of predefined procedures or rules for making
decisions.
Structured Decisions:
• These follow clear, predefined processes or rules and are repetitive.
• They can often be automated due to their fixed nature.
• Example: Approving a loan based on a specific credit score threshold.
Unstructured Decisions:
• These involve unique, complex situations without predefined steps.
• They require creativity, judgment, and experience.
• Example: Deciding how to respond to a sudden data breach.
Semi-structured Decisions:
• These combine elements of both structured and unstructured decisions.
• Some parts are well-defined, while others rely on judgment.
• Example: Allocating a marketing budget based on sales data and creative
ideas.
Strategic Decisions:
• Long-term decisions that affect the entire organization.
• These shape goals, policies, and future direction.
• Example: Expanding the company into a new country.
Tactical Decisions:
• These are medium-term decisions that impact specific areas or
departments.
• They support the organization’s broader strategic goals.
• Example: Launching a regional advertising campaign to support
expansion.
Operational Decisions:
• Short-term, routine decisions that deal with daily activities.
• These ensure smooth functioning of operations.
• Example: Scheduling employees for a weekly work shift.
Q4. Development of a decision support system
DSS: A decision support system (DSS) is an interactive computer-based
application that combines data and mathematical models to help decision
makers solve complex problems faced in managing the public and private
enterprises and organizations.
Planning:
• Identify the goals and needs for building the DSS and assess its feasibility.
• Conduct a cost-benefit analysis to determine its value.
• Example: A company analyzes whether investing in a DSS to improve
supply chain efficiency is worthwhile.
Analysis:
• Define the DSS's functions, target users, and the decision making
processes it will support.
• Ensure it aligns with the organizational environment.
• Example: The company decides the DSS should predict inventory
shortages and notify the warehouse team.
Design:
• Plan the structure and working of the DSS, including hardware, software,
and database requirements.
• Focus on creating a system that’s user-friendly and efficient.
• Example: The company chooses cloud servers, AI models, and a simple
interface for warehouse staff.
Implementation:
• Build and test the DSS step by step using rapid prototyping to minimize
risks. • Gather user feedback and refine the system as needed.
• Example: The company tests a prototype for tracking one product's
inventory before scaling it for all products.
Delivery:
• Deploy the finished DSS, ensuring it meets organizational goals and
satisfies users.
• Provide training and ongoing support if required.
• Example: The DSS is fully operational, offering real-time inventory
predictions and detailed reports to decision-makers.
Q5 define System explain closed and open system.
Definition of a System
• A system is a set of interconnected components that work together to
achieve a specific goal.
• It consists of input, processing, and output components and may include
feedback mechanisms to regulate its performance.
• Systems can be physical (e.g., a manufacturing plant) or conceptual (e.g.,
a decision-making framework).
• In the context of business and marketing, a system helps manage
operations efficiently by structuring processes such as production, sales,
customer engagement, and market analysis.
Closed Cycle Marketing System
• A closed cycle marketing system is a self-contained marketing process
where feedback from previous marketing activities directly influences
future marketing decisions.
• The system continuously refines itself based on the collected data
without relying on external influences.
Characteristics of a Closed Cycle Marketing System:
• Operates within a fixed environment with limited external interference.
• Uses internal feedback loops to improve marketing strategies.
• Often focuses on existing customers rather than reaching new markets.
Example of a Closed Cycle Marketing System:
• A subscription-based streaming service (e.g., Netflix, Spotify) collects
data on user preferences and viewing/listening habits.
• This data is then used to recommend personalized content and improve
customer retention.
• The company does not rely heavily on external data but rather on
internal user behavior analysis to enhance its marketing strategies.
Open Cycle Marketing System
• An open cycle marketing system incorporates external factors such as
market trends, competition, and customer feedback to continuously
adapt its marketing strategies.
• It is a dynamic system that evolves based on external inputs.
Characteristics of an Open Cycle Marketing System:
• Incorporates external market influences such as customer behavior,
competitor strategies, and economic changes.
• Requires continuous data collection from external sources.
• Adapts and changes marketing strategies based on new insights from
the market.
Example of an Open Cycle Marketing System:
A fast-food chain (e.g., McDonald’s) operates in an open cycle marketing
system. It gathers information from market trends, customer feedback, and
competitor pricing to adjust its menu, pricing strategies, and promotional
campaigns.
Unit 2:
Q1 Math Model
Def :
• Explanation: The mathematical model is an abstract model which uses
language to describe the behavior of the system.
• Role of a mathematical model in business intelligence is to understand
the function of systems.
• The mathematical modeling is classified into a black box or a white box
model
Phases in dev of Math Model:
1. Problem Identification
This is the first and most critical phase, where the problem is clearly defined,
and the objectives of the model are established.
Key Activities in this Phase:
• Understanding the real-world problem that needs to be solved.
• Identifying the key variables that influence the problem.
• Defining the scope and limitations of the model.
2. Model Formulation
In this phase, the problem is translated into a mathematical representation.
This involves defining variables, parameters, constraints, and relationships
among them.
Key Components of Model Formulation:
• Decision Variables: Represent the choices available (e.g., the number of
products to stock).
• Objective Function: Defines what needs to be optimized (e.g., minimize
inventory costs).
• Constraints: Limits or restrictions on decision variables (e.g., warehouse
capacity).
• Mathematical Relationships: Equations and inequalities that define how
variables interact.
3. Development of Algorithm
Once the model is formulated, a solution method is chosen. This phase involves
selecting or developing an algorithm that can compute the optimal or near-
optimal solution.
Key Activities in this Phase:
• Selecting the appropriate solution technique
• Developing an algorithm or using existing software tools to solve the
model.
• Running initial simulations to check the feasibility of the solution.
4. Implementation and Testing
In this phase, the model is tested in real-world scenarios to ensure its accuracy
and reliability. If necessary, adjustments are made to improve performance.
Key Activities in this Phase:
• Data validation: Ensuring input data is accurate.
• Model testing: Running different scenarios to check if the model
produces realistic results.
• Implementation: Integrating the model into the decision-making
process.
• Performance evaluation: Comparing the model's predictions with actual
outcomes.
Feedback Loop and Continuous Improvement
Since business environments and data trends change over time, mathematical
models may require modifications. A feedback loop is essential for continuous
improvement.
Steps in the Feedback Loop:
1. Analyze Results
2. Identify Errors
3. Refine the Model
4. Re-test and Validate
Q2 Data Mining Application with example
1 Relational Marketing:
• Using data to understand and market to customers more effectively.
• Examples:
• Identify which customers are likely to buy a new product (cross-selling).
Predict how many people will respond to a discount email.
• Analyze what products customers often buy together (e.g., bread and
butter).
2 Fraud Detection:
• Finding unusual patterns that might indicate fraud.
• Examples: Detecting stolen credit card use by noticing sudden, large
purchases in unfamiliar locations.
• Identifying fake insurance claims by analyzing claim patterns.
3 Risk Evaluation:
• Estimating potential risks before making decisions.
• Example: A bank uses data to decide if someone applying for a loan is
likely to repay it based on their income and past financial behavior.
4 Text Mining:
• Analyzing text data to categorize or extract information.
• Examples: A search engine organizes and ranks web pages based
on their content.
• Email services filter out spam messages automatically.
5 Image Recognition:
• Analyzing and classifying digital images.
• Examples:
• A phone unlocking feature that recognizes your face.
• Sorting pictures in a gallery by identifying objects like "beach" or
"mountains."
6 Web Mining:
• Understanding user behavior on websites.
• Examples: Analyzing which pages users visit before making a
purchase.
• Showing personalized ads based on what a user clicks on a
website.
7 Medical Diagnosis:
• Using data to detect and predict diseases early
• Examples:
• Identifying cancer in scans using AI.
• Predicting diabetes risk based on test results and patient history.
Q3 supervised and unsupervised learning
Supervised and Unsupervised Learning in Data Mining
• Supervised learning and unsupervised learning are two primary
approaches in machine learning and data mining.
• Both methods help in analyzing data patterns, making predictions, and
gaining insights, but they differ in how they process and interpret data.
1. Supervised Learning
Definition
• Supervised learning is a machine learning technique where the model is
trained using labeled data.
• Each input data point is associated with a corresponding output (label),
and the algorithm learns to map inputs to the correct outputs.
Key Characteristics:
• Requires labeled data (i.e., data with known outcomes).
• Learns from examples and generalizes to unseen data.
• Used for prediction and classification tasks.
Types of Supervised Learning:
1. Classification:
o The output is a category or label.
o Example: Spam detection in emails (Spam / Not Spam).
2. Regression:
o The output is a continuous numerical value.
o Example: Predicting house prices based on location and size.
Examples of Supervised Learning:
• Fraud detection
• Medical diagnosis
• Speech recognition
• Stock price prediction
2. Unsupervised Learning
Definition
• Unsupervised learning is a machine learning technique where the model
is trained using unlabeled data.
• The algorithm identifies hidden patterns, structures, or relationships in
the data without prior knowledge of the outcomes.
Key Characteristics:
• Works with unlabeled data (i.e., no predefined outputs).
• Learns underlying structures in data without explicit instructions.
• Used for clustering and association tasks.
Types of Unsupervised Learning:
1. Clustering:
o The algorithm groups similar data points together.
o Example: Grouping customers based on purchasing behavior.
2. Association Rule Learning:
o Finds relationships between different data points.
o Example: Market basket analysis (people who buy bread often buy
butter).
Examples of Unsupervised Learning:
• Customer segmentation
• Anomaly detection
• Recommendation systems
• Genetic clustering
Q4 data validation
Causes of Poor Data Quality
The quality of input data may prove unsatisfactory due to the following
reasons:
1. Incompleteness: Some records may have missing values due to various
reasons such as data not being recorded systematically, unavailable at
the time of transaction, or lost during data transfer.
2. Noise: Data may contain errors or anomalies (outliers), often due to
malfunctioning measurement devices, recording errors, or differences in
measurement units that require conversion.
3. Inconsistency: Discrepancies in data can arise due to changes in coding
systems, data integration issues, or lack of standardization across
different periods.
Data Validation and Its Importance
Definition:
Data validation is a process used to identify and correct errors in data to ensure
its accuracy, completeness, and consistency before being used for analysis or
decision-making. The main purpose of validation is to prevent issues caused by
incomplete, noisy, or inconsistent data.
Techniques for Data Validation
1. Elimination:
o Discarding records with missing values.
o Ensures accuracy but may lead to loss of significant data.
2. Inspection:
o Experts manually review missing values to suggest replacements.
o Effective but time-consuming for large datasets.
3. Identification and Encoding:
o Assigning a placeholder value to missing data instead of deleting
records.
o Example: Using -1 for missing numerical values.
4. Substitution:
o Replacing missing values with estimated values like mean, median,
or mode.
o Can be done using statistical methods or machine learning
techniques.
Importance of Data Validation
1. Ensures Data Accuracy: Prevents incorrect data from being processed.
2. Improves Decision-Making: Ensures reliable data for business
intelligence.
3. Maintains Data Integrity: Reduces inconsistencies in datasets.
4. Enhances Efficiency: Reduces time spent on data correction and
cleaning.
5. Prevents Costly Errors: Avoids financial and operational issues due to
bad data.
By applying data validation techniques, organizations can enhance the
reliability and usability of their data, making it suitable for business intelligence
and data mining applications.
Q5 Predictive and Optimization model
Predictive and Optimization Models
• Mathematical models play a crucial role in business intelligence by
supporting data-driven decision-making.
• Among these, predictive models and optimization models are two of
the most widely used approaches for forecasting trends and improving
efficiency.
1. Predictive Models
Definition
• Predictive models play a primary role in business intelligence systems
since they are logically placed upstream in the decision-making process.
• These models use historical data and statistical techniques to forecast
future events.
Key Features:
• Used in strategy, research and development, marketing, production,
and logistics.
• Employs pattern recognition, statistical analysis, and machine learning
techniques.
• Helps organizations anticipate trends and risks to improve decision-
making.
Applications of Predictive Models:
1. Customer Behavior Analysis:
o Predicts customer preferences and buying habits.
o Example: E-commerce platforms suggest products based on past
purchases.
2. Risk Assessment:
o Evaluates potential financial or operational risks.
o Example: Banks assess loan applicants' risk profiles before
approving loans.
3. Sales and Demand Forecasting:
o Predicts future demand to optimize inventory and production.
o Example: A retail company estimates holiday season sales to stock
up in advance.
2. Optimization Models
Definition
• Optimization models aim to find the best possible solution to a decision-
making problem by maximizing or minimizing an objective function (e.g.,
profit maximization, cost minimization).
• These models are used when limited resources need to be allocated
effectively among competing needs.
Key Features:
• Defines a set of feasible decisions based on constraints and conditions.
• Establishes evaluation criteria to compare alternative choices.
• Helps decision-makers identify the most efficient and cost-effective
solutions.
Applications of Optimization Models:
1. Supply Chain Management:
o Minimizes transportation and inventory costs while ensuring
timely deliveries.
o Example: A logistics company optimizes delivery routes to reduce
fuel consumption.
2. Production Planning:
o Allocates resources efficiently in manufacturing processes.
o Example: A factory schedules machine usage to meet production
targets with minimal downtime.
3. Workforce Scheduling:
o Assigns employees to shifts in a way that maximizes productivity
while reducing labor costs.
o Example: A call center assigns workers to peak hours based on
customer demand.
Unit 3;
Taxonomy and Phases of Classification Models
1. Taxonomy of Classification Models
Classification models are supervised learning techniques used to predict the
categorical target feature of new instances based on historical data. They can
be categorized into four main types:
1.1 Probabilistic Models
These models estimate the probability that an observation belongs to a specific
class based on prior knowledge.
• Naïve Bayes Classifier: Assumes conditional independence between
features and applies Bayes' theorem to classify instances.
• Bayesian Networks: Utilize probabilistic graphical models to represent
dependencies between variables.
1.2 Linear Models
Linear models create decision boundaries using linear equations.
• Logistic Regression: Estimates the probability of an instance belonging to
a class using a sigmoid function.
• Support Vector Machines (SVM): Identifies an optimal hyperplane to
separate different classes.
1.3 Tree-Based Models
These models use decision trees to classify instances by splitting the dataset
based on feature values.
• Decision Trees: Uses a hierarchical structure to classify data points.
• Random Forest: An ensemble learning method that improves
classification accuracy by combining multiple decision trees.
1.4 Neural Networks
Inspired by biological neurons, these models use multiple layers to process
input data and make predictions.
• Multilayer Perceptron (MLP): A deep learning model with multiple
hidden layers.
• Convolutional Neural Networks (CNN): Specialized for image
classification tasks.
2. Phases of Classification Models
The development of a classification model follows three main phases:
2.1 Training Phase
• The model learns from a labeled dataset, known as the training set (T).
• The classification algorithm identifies patterns and generates
classification rules.
• Example: A spam email classifier learns from past labeled emails.
2.2 Test Phase
• The model is evaluated using a separate test set (V = D - T).
• Predicted classifications are compared with actual values to measure
accuracy.
• Ensures that the model generalizes well to unseen data.
2.3 Prediction Phase
• The trained model is applied to new, unseen data.
• Uses the classification rules learned in the training phase to assign labels.
• Example: A loan approval system classifies applicants as "Approved" or
"Rejected" based on financial history.
Q2 K- means algo for clustering
K-Means Algorithm for Clustering
• Clustering is an unsupervised learning technique used to group similar
data points into clusters.
• It helps in identifying patterns and relationships in large datasets.
• One of the most widely used clustering techniques is the K-Means
algorithm.
• The K-Means algorithm is a partition-based clustering technique that
divides a dataset into K clusters, where K is a predefined number.
• It aims to minimize the variance within each cluster while maximizing
the difference between clusters.
Steps in the K-Means Algorithm
1. Initialization:
o Select K random points from the dataset as initial cluster
centroids.
2. Assignment of Data Points:
o Each data point is assigned to the nearest cluster centroid based
on Euclidean distance.
3. Updating Centroids:
o Compute the new centroid for each cluster by taking the mean of
all data points assigned to it.
4. Iteration and Convergence:
o Repeat step 2 and step 3 until centroids no longer change,
meaning the clusters are stable.
Example of K-Means Clustering
Suppose a retail company wants to segment its customers based on their
purchasing behavior. Using K-Means clustering, the company can divide
customers into groups such as:
• Cluster 1: High-spending frequent buyers.
• Cluster 2: Medium-spending occasional buyers.
• Cluster 3: Low-spending rare buyers.
This helps in targeted marketing and personalized recommendations.
Q3 Partition and Hierarchical Methods with Agglomerative and Divisive
Approaches
• Clustering methods aim to group similar data points into clusters based
on their characteristics.
• Two major approaches to clustering are partition methods and
hierarchical methods.
1. Partition Methods
Definition
Partitioning techniques divide a dataset DDD containing mmm observations
into KKK clusters, where KKK is predefined. The goal is to create homogeneous
clusters where each observation belongs to one cluster exclusively.
Characteristics of Partition Methods
• Requires predefined K clusters.
• Uses an iterative approach to minimize intra-cluster variance.
• Suitable for small to medium-sized datasets.
• Does not form a hierarchical structure.
Examples of Partitioning Methods
1. K-Means Algorithm
o Uses centroids to define clusters.
o Minimizes the sum of squared distances between points and
cluster centers.
2. K-Medoids Algorithm
o Similar to K-Means but uses medoids (actual data points) instead
of centroids.
o More robust to outliers compared to K-Means.
2. Hierarchical Methods
Definition
Hierarchical clustering methods create a tree-like structure (dendrogram),
representing nested clusters. These techniques do not require predefining the
number of clusters.
Characteristics of Hierarchical Methods
• No need to predefine the number of clusters.
• Generates a hierarchical representation of data.
• Suitable for datasets of varying sizes.
• Uses linkage methods (e.g., minimum, maximum, mean, centroid,
Ward’s method) to determine distances between clusters.
Agglomerative vs. Divisive Hierarchical Methods
Hierarchical clustering techniques can be classified into two types:
3.1 Agglomerative (Bottom-Up) Method
• Starts with each observation as its own cluster.
• Iteratively merges the closest clusters until only one remains.
• Uses different linkage criteria to measure distances between clusters.
Example:
A customer segmentation model where each customer starts in an individual
cluster, and similar ones are merged based on their spending habits.
3.2 Divisive (Top-Down) Method
• Starts with all observations in a single cluster.
• Iteratively splits clusters into smaller sub-clusters.
• Continues until each observation is its own cluster or meets a predefined
threshold.
Example:
A document classification system where all documents start in one cluster and
are split based on keyword similarity.
Q4 Confusion Matrix
• A confusion matrix is a table used to evaluate the performance of a
classification model.
• It compares the predicted classifications with the actual values, helping
in analyzing how well the model distinguishes between different classes.
• The confusion matrix is a crucial tool in classification model evaluation.
• It provides insights into model performance beyond simple accuracy,
helping businesses and researchers optimize predictive models for real-
world applications.
Structure of a Confusion Matrix
For a binary classification problem, the confusion matrix is a 2×2 table as
shown below:
Actual / Predicted Predicted Positive Predicted Negative
Actual Positive True Positive (TP) False Negative (FN)
Actual Negative False Positive (FP) True Negative (TN)
• True Positive (TP): Model correctly predicts a positive case.
• False Negative (FN): Model incorrectly predicts a negative case when it
should be positive.
• False Positive (FP): Model incorrectly predicts a positive case when it
should be negative.
• True Negative (TN): Model correctly predicts a negative case.
Measures in Confusion Matrix:
error rate and accuracy are two key performance metrics derived from the
confusion matrix. These measures help evaluate how well a classification
model performs.
1. Accuracy
Definition:
• Accuracy is the proportion of correctly classified instances (both True
Positives (TP) and True Negatives (TN)) out of the total number of
predictions.
• It represents the overall correctness of the model.
2. Error Rate
Definition:
• Error rate is the proportion of incorrect predictions (False Positives (FP)
and False Negatives (FN)) over the total predictions.
• It measures how often the model makes incorrect classifications.
Q5 naïve baysian classification
Naïve Bayesian Classification
• Naïve Bayesian classification is based on Bayes' theorem, which
calculates the probability that a given observation belongs to a specific
class.
• It assumes that the features used for classification are conditionally
independent given the class label.
3. Assumption of Conditional Independence
The "naïve" assumption is that all features are independent of each other
given the class label. This simplifies the computation:
This assumption makes the model computationally efficient, even for large
datasets.
4. Steps in Naïve Bayes Classification
1. Training Phase
o Compute prior probabilities for each class.
o Compute likelihood probabilities for each feature given the class.
2. Prediction Phase
o Use Bayes' theorem to calculate the probability of each class for a
given data point.
o Assign the class with the highest probability.
Extra - Bayesian Classification (General Bayesian Approach)
• Bayesian classification refers to a broad category of classification
techniques based on Bayes' theorem.
• It computes the probability of a class given observed data and can use
complex probabilistic models to handle dependencies between
features.
• Bayesian classifiers do not assume feature independence and can work
with correlated attributes.
• Example: Bayesian networks, which use directed graphs to model
dependencies between features.
Unit 4:
Q1 relational marketing
Definition of Relational Marketing
• Relational marketing is a customer-focused marketing strategy aimed at
building long-term relationships with customers rather than just driving
one-time sales.
• It involves understanding customer needs, personalizing interactions,
and enhancing engagement through targeted marketing efforts.
Customer Lifetime in a Cycle
The lifetime of a customer follows several stages, reflecting their journey with
a business:
1. Prospect Stage: A potential customer is identified through marketing
efforts.
2. Acquisition Phase: The prospect converts into a customer by making a
first purchase or signing up for a service.
3. Revenue Growth Phase: The customer relationship matures through
repeat purchases, cross-selling, and up-selling.
4. Retention Phase: The company focuses on keeping customers engaged
and preventing churn.
5. Exit Phase: The customer discontinues the service or stops buying,
which can happen voluntarily (switching to a competitor), involuntarily
(payment issues), or unintentionally (relocation).
Businesses aim to maximize the customer lifetime value (CLV) by
implementing strategies to increase retention, loyalty, and engagement.
Data Mining Applications in Relational Marketing
Data mining is a powerful tool in relational marketing, helping businesses
analyze customer data for better decision-making. Some key applications
include:
1. Customer Segmentation: Identifies groups of customers with similar
behaviors for targeted campaigns.
2. Predicting Campaign Response: Uses historical data to estimate how
likely customers are to respond to promotions.
3. Understanding Buying Behavior: Analyzes purchase patterns to suggest
relevant products and services.
4. Market Basket Analysis: Identifies products frequently bought together
to optimize cross-selling and bundling strategies.
Motivations for Relational Marketing
Several factors have contributed to the rise of relational marketing:
• Market Complexity: The increase in large enterprises has made
customer interactions more complex.
• Shorter Product Lifecycles: Rapid innovation has led to more frequent
product updates.
• Increased Competition: Globalization and e-commerce have enabled
easy comparison of prices and features.
• Customer Loyalty Challenges: Switching service providers has become
easier.
• Data Availability: Businesses can now leverage vast amounts of
customer data for targeted marketing
Objectives of Relational Marketing
The key objectives of relational marketing include:
1. Customer Retention & Loyalty:
o Focuses on building long-term relationships rather than one-time
sales.
2. Personalized Marketing:
o Uses customer data to provide tailored offers and
communication.
3. Maximizing Customer Lifetime Value (CLV):
o Ensures that customers continue to generate revenue over an
extended period.
4. Optimizing Marketing Efforts:
o Reduces wasted resources by targeting the right audience
effectively.
5. Competitive Advantage:
o Helps companies differentiate themselves in a competitive market
through superior customer engagement.
6. Efficient Cross-Selling & Up-Selling:
o Encourages customers to purchase complementary products or
higher-value alternatives.
Q2 market basket analysis
1. Definition
• Market Basket Analysis (MBA) is a data mining technique used to
discover relationships between items frequently purchased together.
• It helps businesses analyze customer buying patterns to improve cross-
selling, up-selling, and promotional strategies.
2. Key Concepts in Market Basket Analysis
1. Basket (Transaction):
o A set of items purchased together in a single transaction.
o Example: If a customer buys bread and butter, they are part of the
same basket.
2. Association Rules:
o Used to identify relationships between items in transactions.
o Example: If customers buy milk, they are likely to buy cereal.
3. Support, Confidence, and Lift:
o Support: The percentage of transactions containing both items.
o Confidence: The likelihood that if item A is bought, item B will also
be bought.
o Lift: Measures how much the likelihood of buying B increases
when A is purchased.
3. Applications of Market Basket Analysis
1. Retail and E-Commerce:
o Product Placement: Stores arrange products frequently bought
together near each other.
o Recommendation Systems: Amazon and Flipkart suggest items
based on past purchases.
2. Banking & Finance:
o Credit Card Offers: Identifying customer spending patterns to
provide tailored credit card promotions.
3. Telecommunications:
o Service Bundling: Mobile providers suggest complementary plans
based on usage history.
4. Healthcare & Insurance:
o Prescription Patterns: Identifying medications often prescribed
together for better patient care.
Q3 supply chain management
1. Definition
Supply chain optimization refers to the strategic planning and operational
improvements that enhance efficiency, reduce costs, and ensure a smooth flow
of goods, services, and information from suppliers to end customers.
2. Objectives of Supply Chain Optimization
• Minimizing Total Costs: Reducing procurement, production, inventory,
and transportation costs.
• Maximizing Service Levels: Ensuring timely product delivery and
customer satisfaction.
• Efficient Resource Utilization: Optimizing manufacturing, warehousing,
and logistics.
• Improving Demand Forecasting: Using predictive analytics to align
supply with customer demand.
3. Optimization Models for Supply Chains
1. Linear Programming (LP):
o Optimizes resource allocation for manufacturing and logistics.
o Example: Minimizing transportation costs while meeting delivery
deadlines.
2. Inventory Optimization Models:
o Determines the optimal stock levels based on demand patterns.
o Example: Economic Order Quantity (EOQ) model.
3. Network Optimization:
o Optimizes the distribution network by selecting the best
warehouses, suppliers, and delivery routes.
o Example: Reducing the distance between warehouses and
customers.
Q4 CCR Model
CCR Model (Charnes-Cooper-Rhodes Model)
1. Definition
The CCR Model (Charnes-Cooper-Rhodes Model) is a Data Envelopment
Analysis (DEA) approach used to evaluate the efficiency of decision-making
units (DMUs), such as businesses, hospitals, or bank branches. It measures
how well a unit converts inputs (e.g., labor, capital) into outputs (e.g., products,
services).
2. Key Features of the CCR Model
1. Linear Optimization Model:
o The CCR model uses linear programming to compute efficiency
scores.
2. Efficiency Score Calculation:
The score ranges from 0 to 1, where 1 indicates an efficient unit.
3. Input-Oriented & Output-Oriented Approaches:
o Input-Oriented: Minimizes inputs while maintaining the same
output.
o Output-Oriented: Maximizes output with the same level of inputs.
4. Efficient Frontier:
o The most efficient DMUs form the efficient frontier, against which
all other DMUs are compared.
3. Dual CCR Model
The dual model of CCR identifies an "ideal unit" that serves as a benchmark.
This helps organizations:
✔ Identify inefficient units.
✔ Set target improvement levels.
✔ Optimize resource allocation.
4. Applications of the CCR Model
• Banking: Evaluating branch performance based on loans and customer
accounts.
• Healthcare: Comparing hospitals based on patient care efficiency.
• Education: Measuring university efficiency in research output versus
funding.
Q5 tactical/optimization model for logistic planning
Tactical/Optimization Model for Logistic Planning
1. Introduction
Tactical and optimization models for logistic planning aim to enhance the
efficiency of supply chain operations by reducing costs while ensuring timely
delivery of goods and services. These models help businesses align production,
inventory, and distribution with market demand.
2. Objectives of Logistic Planning
• Minimizing Total Costs: Covers procurement, processing, inventory, and
transportation costs.
• Maximizing Resource Utilization: Efficiently using production capacity
and warehouse space.
• Ensuring Timely Deliveries: Reducing delays in supply chains.
• Balancing Supply and Demand: Matching production levels with
customer needs.
3. Optimization Models for Logistics Planning
1. Integrated Supply Chain Optimization Model
o Analyzes the entire logistics network, including suppliers,
manufacturers, and retailers.
o Reduces overall costs while maintaining service levels.
2. Inventory Optimization Model
o Determines the optimal stock levels for warehouses and stores.
o Prevents both overstocking (waste) and understocking (lost
sales).
3. Transportation Optimization Model
o Focuses on route planning and fleet management to minimize
transportation costs.
o Uses linear programming and heuristic algorithms for efficiency.
4. Multiple Resource Constraint Model
o Accounts for limited availability of manufacturing resources
(labor, equipment, raw materials).
o Ensures balanced production scheduling across different facilities.
5. Backlogging and Demand Forecasting Model
o Backlogging: Allows fulfilling delayed orders in a future period to
reduce shortages.
o Forecasting: Uses AI and statistical methods to predict future
demand trends.
4. Tactical Planning in Logistics
• Medium-Term Planning: Covers monthly or quarterly adjustments in
logistics operations.
• Production Volume Optimization: Determines how much of each
product to manufacture in a given period.
• Decision Variables: Involves capacity allocation, shipment scheduling,
and supplier selection.
Unit 5:
Q1 whats KMS and explain KMS cycle
What is KMS?
A Knowledge Management System (KMS) is a system designed to facilitate the
creation, organization, sharing, and management of knowledge within an
organization. It transforms information into knowledge and ensures its
accessibility for decision-making, learning, and innovation.
KMS Cycle
The Knowledge Management System (KMS) Cycle consists of six steps that
define how knowledge is created, processed, stored, and shared within an
organization:
1. Creating Knowledge – Knowledge is generated by individuals through
sharing insights, ideas, and experiences within an organization.
2. Capturing Knowledge – The generated knowledge is collected and
stored in a knowledge repository for future reference.
3. Refining Knowledge – The captured knowledge is structured and
organized using a framework or knowledge model.
4. Storing Knowledge – The refined knowledge is stored in a systematic
format within a knowledge repository for easy retrieval.
5. Managing Knowledge – The knowledge is maintained and made
searchable, ensuring its relevance and usability.
6. Disseminating Knowledge – The stored knowledge is shared across the
organization so that employees can access and apply it when needed.
Q2 diff between process and practice approach in KMS
Difference Between Process and Practice Approach in KMS
In Knowledge Management Systems (KMS), two primary approaches exist for
managing knowledge:
1. Process Approach
o Focuses on structured processes, controls, and technology to
codify and manage knowledge.
o Organizations using this approach implement explicit rules for
knowledge collection, storage, and sharing.
o Suitable for environments where knowledge can be easily
documented, stored, and retrieved (e.g., databases, manuals).
o Often relies on Information and Communication Technology (ICT)
to facilitate knowledge dissemination.
2. Practice Approach
o Assumes that organizational knowledge is tacit (experiential) and
not easily codified.
o Rejects formal controls and structured processes, emphasizing
social interaction and communities of practice.
o Encourages peer-to-peer learning, mentorship, and informal
discussions for knowledge sharing.
o Best for situations where knowledge is complex and context-
dependent, requiring direct experience to understand.
Most organizations adopt a Hybrid Approach, combining both process and
practice elements for a balanced Knowledge Management strategy
Q3 Knowledge eng
Knowledge Engineering
Introduction
Knowledge Engineering (KE) is a discipline within Artificial Intelligence (AI) that
focuses on creating systems capable of mimicking human decision-making. It
involves the acquisition, representation, and application of knowledge to solve
complex problems efficiently. KE plays a crucial role in the development of
Expert Systems, which are designed to simulate human reasoning and provide
solutions in specific domains like healthcare, finance, and engineering.
Components of Knowledge Engineering
1. Knowledge Base
o A repository of structured information containing both factual
knowledge (universally accepted information) and heuristic
knowledge (expert judgment and best practices).
o Knowledge Representation methods, such as IF-THEN-ELSE rules,
are used to store and process this knowledge.
2. Knowledge Acquisition
o The process of gathering expert knowledge through interviews,
case studies, and observation.
o Knowledge engineers interact with domain experts to translate
their expertise into a structured format for computational use.
3. Inference Engine
o The "brain" of an expert system, which applies logical reasoning to
derive new knowledge from existing data.
o Uses Forward Chaining (predictive approach) or Backward
Chaining (diagnostic approach) to draw conclusions.
4. User Interface
o Enables users to interact with the system, input data, and receive
outputs in a comprehensible manner.
Q4 whats expert system with its application, how is it diff from DSS
Expert System and Its Applications
Introduction
An Expert System (ES) is a type of Artificial Intelligence (AI) that mimics human
decision-making by using a structured knowledge base and inference engine.
These systems provide expert-level solutions to problems in specialized
domains such as medicine, engineering, finance, and business analytics.
Unlike general computing systems, Expert Systems do not rely solely on stored
data; they apply logical reasoning and predefined rules to derive conclusions.
Components of an Expert System
1. Knowledge Base – Stores facts, rules, and heuristics (expert knowledge).
2. Inference Engine – Applies reasoning to draw conclusions from the
knowledge base.
3. User Interface – Allows interaction between the system and the user.
Applications of Expert Systems
Expert Systems are widely used in various industries for problem-solving and
decision-making. Some common applications include:
• Medical Diagnosis – Systems like MYCIN assist doctors in diagnosing
bacterial infections.
• Engineering Troubleshooting – Diagnosing faults in machinery, power
grids, and software systems.
• Finance and Banking – Fraud detection, risk assessment, and loan
approval.
• Agriculture – Soil analysis, pest detection, and crop disease prediction.
• Education – Intelligent tutoring systems that guide students through
learning paths.
Q5 explain in brief: -
Comparison: Artificial Intelligence (AI) vs. Natural Intelligence (NI)
Artificial Intelligence (AI)
• AI systems process large amounts of data at high speed using predefined
algorithms.
• AI is specialized; a model trained for one task (e.g., image recognition)
cannot automatically perform another (e.g., language translation).
• AI does not have self-awareness or emotions; it follows logical decision-
making patterns.
• AI is highly scalable—it can be replicated easily across systems and
environments.
• AI can operate 24/7 without fatigue, making it efficient for repetitive
tasks.
• AI lacks general intelligence—it performs well in structured
environments but struggles in unpredictable situations.
Natural Intelligence (NI)
• NI allows humans to adapt to new and unknown situations through
reasoning and intuition.
• NI involves emotional intelligence, enabling empathy, creativity, and
ethical decision-making.
• NI learns from experience rather than requiring explicit programming.
• NI allows for multi-tasking, whereas AI models are usually trained for
specific tasks.
• NI includes consciousness and self-awareness, leading to personal
growth and innovation.
• NI is resource-efficient—the human brain consumes very little energy
compared to AI supercomputers
Difference Between Data, Information, and Knowledge
1. Data
o Definition: Raw facts, figures, or measurements without context.
o Example: "500", "Mumbai", "Temperature: 32°C".
o Nature: Unprocessed and meaningless until analyzed.
2. Information
o Definition: Processed or structured data that provides meaning.
o Example: "The temperature in Mumbai is 32°C today."
o Nature: Organized data that helps in understanding.
3. Knowledge
o Definition: Information that is contextual, relevant, and
actionable.
o Example: "A temperature of 32°C in Mumbai during March
suggests summer has started."
o Nature: Derived from experience, insight, and expertise.
Knowledge is applied information, while information is processed data.