11-130 Org
11-130 Org
Page | 12
a. Understanding Information Gain and Entropy
b. Applications in decision trees and classification
5. Confusion Matrix
a. Understanding metrics like Accuracy, Precision, Recall, FScore
b. Confusion Matrix in real world problems
Page | 13
b. Industrial Use of Naive Bayes (PIMA Diabetic Test)
4. Support Vector Machine (SVM)
a. SVM theory and application
b. Nonlinear SVM, SVM usecase examples
Reinforcement Learning
1. Markov’s Decision Process (MDP)
a. Understanding MDP and its importance in RL
b. QLearning overview and implementation
2. The Bellman Equation
a. Indepth understanding of Bellman Equation
b. Transitioning to QLearning in practical applications
3. Implementing QLearning
a. Implement QLearning in a simulated environment
b. CounterStrike example and its usecase in RL
4. Reinforcement Learning Projects
Page | 14
a. Handson practice with an RL project
b. Apply QLearning to a selfplay game
Python Basics:
1. Introduction
a. I/O statements in python
b. Variables, identifiers, statements, conditions
c. programs
2. Operators
a. Data types
b. Functions/ File handling
c. comprehensions
3. OOPS concepts
a. OOPS basics(class,object, concepts)
b. Programs
Page | 15
2. Real World ML Project
a. Start a complete end to end project Data preprocessing, Model training, and
Evaluation
3. Optimization & Hyperparameter Tuning
a. Tuning hyperparameters using GridSearch and RandomSearch
b. Cross validation in practice
4. Project Demo & Presentation
a. Prepare the project for presentation
b. Evaluate the model on the test set and validate predictions
Page | 16
CHAPTER1: EXECUTIVE SUMMARY
The internship report shall have a brief executive summary. It shall include five or more
Learning Objectives and Outcomes achieved, a brief description of the sector of business
and intern organization and summary of all the activities done by the intern during the
period.
Suggestive contents
Introduction of the Organization Epro academy & Glossary Softech committed EdTech
Organizations with high quality Trainings and Placements record. Epro academy & Glossary
Softech are Involved in training and placements of Engineering, Degree, Polytechnic, ITI,
BPO/Tech Support
Data Management
Software Development
Epro academy & Glossary Softech have a vision to become a INDIA’S Best Training
Organizations and make the students expert in both academics and technology related knowledge and
mission to provide the best quality training at a reasonable price like a non-profitable organization we
follow quality training and value-added services for the students and Working People.
The policy of Epro Academy & Glossary Softech towards of the Interns is to Provide
Best quality Training in Both Online and Offline mode so Intern can get best Job
opportunities in this competitive World.
Page | 17
C. Organizational Structure.
Epro academy and Glossary soft tech have Biggest Lab and Company for Complete
Practical Infrastructures and placements and Trainings.
At Epro academy & Glossary soft tech Organizations we Train on multiple things like IT
trainings, HR & corporate activities, Job roles, Personality Development, Interview preparation
employees take at most importance in providing quality trainings on future trends.
Performance of the Organization in terms of turnover, profits, market reach and market value.
Glossary Soft Tech put emphasis on long-term commitment and combine global reach and
local intimacy to provide premier professional services from consulting, system development to
business IT outsourcing. We share the aspiration of our customers to realize innovation. Epro
academy is the Top most Training and Placement Organization in its region. We have played key
role in Training Engineering, polytechnic, Degree, ITI, Core fields, IT Sector and Medical sector
with excellent growth in both Turnover and quality training.
The Major future plans of Epro academy & Glossary Softech to Establish Skill Oriented
Center of Excellences all over the INDIA hence to involve in nation building.
At Epro academy & Glossary Soft tech Trainee will learn about the selected domain in both Theory
and Particles. live practical knowledge is given to the trainee with industry experts. Trainee will be
Trained by Real time Industry experts and hands on practice will be given. Internship Includes Job
interview preparation and Industry ready Subject Knowledge
Page | 18
ACTIVITY LOG FOR THE FIRST WEEK
Day Person
Brief description of the daily Learning Outcome In-Charge
activity Signature
Page | 19
WEEKLY REPORT
WEEK–1(From Dt………..…..to Dt.................. )
Detailed Report:
Page | 20
Introduction to Machine Learning
Machine Learning (ML) is a subset of Artificial Intelligence (AI) that enables computers to
learn patterns from data and make decisions or predictions without explicit programming. The
core idea is to build models that improve automatically through experience.
AI vs ML vs Deep Learning
Artificial Intelligence (AI): A broad field that involves creating intelligent machines
capable of simulating human thinking.
Machine Learning (ML): A subset of AI where machines learn from data without
being explicitly programmed.
Deep Learning (DL): A further subset of ML that uses neural networks with multiple
layers to analyze complex data.
Supervised Learning: Uses labeled data for training (e.g., spam detection).
Unsupervised Learning: Finds hidden patterns in unlabeled data (e.g., customer
segmentation).
Reinforcement Learning: Trains an agent through rewards and penalties (e.g., game
playing AI).
Page | 21
Supervised Learning
Supervised Learning is a type of ML where the model learns from labeled datasets. Each
input data point has a corresponding correct output, and the algorithm maps the input to the
output.
Classification: Categorizing data into predefined classes (e.g., email spam detection,
disease diagnosis).
Regression: Predicting continuous values (e.g., house price prediction, stock market
forecasting).
Unsupervised Learning
Reinforcement Learning
Games: AI playing chess, Go, or video games (e.g., AlphaGo, OpenAI's Dota 2).
Robotics: Training robots to perform complex tasks (e.g., self-driving cars, industrial
automation).
Page | 22
ACTIVITY LOG FOR THE SECOND WEEK
Day Person
Brief description of the daily Learning Outcome In-Charge
activity Signature
Page | 23
WEEKLY REPORT
WEEK–2 (From Dt………..…..to Dt................. )
Detailed Report:
Page | 24
Setting up Jupyter Notebook
Jupyter Notebook is an open-source interactive computing environment that allows users to create
and share documents containing live code, equations, visualizations, and text. It is widely used in
Machine Learning (ML) for prototyping and experimentation.
Installation Steps:
jupyter notebook
import numpy as np
a = np.array([1, 2, 3])
print(a)
import pandas as pd
df = pd.read_csv("data.csv")
print(df.head())
Page | 25
4. Scikit-learn – For implementing ML algorithms.
Classification is a Supervised Learning technique where the goal is to predict categorical labels.
Anomaly detection is used for identifying rare events or unusual patterns in data.
Introduction to Clustering
Clustering is an Unsupervised Learning technique where similar data points are grouped together
based on features.
1. K-Means Clustering – Groups data into ‘K’ clusters based on the centroid approach.
Page | 26
Regression Algorithms
Introduction to Regression
1. Linear Regression:
o Predicts continuous values.
o Example: Predicting house prices based on area.
o Formula: Y=mX+bY = mX + bY=mX+b
2. Logistic Regression:
o Used for binary classification problems.
o Example: Spam detection (Spam or Not Spam).
o Uses a sigmoid function to map outputs between 0 and 1.
The Iris dataset contains 150 samples of iris flowers, categorized into three species: Setosa,
Versicolor, and Virginica.
Features include sepal length, sepal width, petal length, and petal width.
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
from sklearn import datasets
from sklearn.model_selection import train_test_split
from sklearn.preprocessing import StandardScaler
from sklearn.neighbors import KNeighborsClassifier
from sklearn.metrics import accuracy_score, confusion_matrix
iris = datasets.load_iris()
X = iris.data
Page | 27
y = iris.target
scaler = StandardScaler()
X_train = scaler.fit_transform(X_train)
X_test = scaler.transform(X_test)
model = KNeighborsClassifier(n_neighbors=3)
model.fit(X_train, y_train)
y_pred = model.predict(X_test)
accuracy = accuracy_score(y_test, y_pred)
print(f'Accuracy: {accuracy * 100:.2f}%')
# Confusion Matrix
conf_matrix = confusion_matrix(y_test, y_pred)
sns.heatmap(conf_matrix, annot=True, cmap="Blues", fmt="d")
plt.xlabel("Predicted")
plt.ylabel("Actual")
plt.show()
Key Takeaways:
Summary:
✓
⬛ Setting up and using Jupyter Notebook for ML development.
✓
⬛ Introduction to classification, clustering, and regression algorithms.
✓
⬛ Hands-on experience with Scikit-learn using the Iris dataset.
Page | 28
ACTIVITY LOG FOR THE THIRDWEEK
Day Person
Brief description of the daily Learning Outcome In-Charge
activity Signature
Page | 29
WEEKLY REPORT
WEEK–3(From Dt………..…..to Dt.................. )
Detailed Report:
Page | 30
Statistics & Probability for Machine Learning
Categories of Data
Introduction to Statistics
1. Population vs Sample:
o Population: Entire dataset or group under study (e.g., all students in a country).
o Sample: Subset of the population used for analysis (e.g., students in a particular
city).
2. Parameter vs Statistic:
o Parameter: A value that describes a population (e.g., average height of all
students).
o Statistic: A value derived from a sample (e.g., average height of students in one
school).
3. Variables:
o Independent Variable (Predictor): The input variable influencing the outcome.
Page | 31
oDependent Variable (Response): The outcome being measured.
4. Bias & Variability:
o Bias: Systematic error in data leading to inaccurate results.
o Variability: Degree of spread in data points (higher variability = less reliable
predictions).
Sampling Techniques
1. Random Sampling:
o Every individual has an equal chance of selection.
o Example: Picking names randomly from a list.
2. Systematic Sampling:
o Selecting every ‘k-th’ individual from a list.
o Example: Selecting every 10th customer in a survey.
3. Stratified Sampling:
o Population is divided into subgroups (strata), and random samples are taken from
each.
o Example: Sampling equal proportions from male and female students in a university.
Importance of Sampling:
Descriptive Statistics
1. Mean (Average):
Mean= (∑X)/N
Page | 32
o Example: The most frequently purchased mobile phone model.
1. Variance (σ²):
o Measures how far data points are from the mean.
Variance=∑(X−Mean)2/N
import numpy as np
Probability Theory
Probability is used to measure the likelihood of events occurring in data analysis and machine
learning.
1. Independent Events:
o The occurrence of one event does not affect the probability of another.
o Example: Rolling a die and flipping a coin.
2. Dependent Events:
o The probability of one event depends on another.
o Example: Drawing cards without replacement from a deck.
Page | 33
3. Mutually Exclusive Events:
o Two events cannot happen at the same time.
o Example: A single dice roll cannot be both 1 and 6.
4. Conditional Probability:
o The probability of event A occurring given that event B has already occurred.
P(A∣B)=P(A∩B)P(B)
o Example: Probability of drawing a King given that a Queen was drawn earlier from
a deck.
Probability Distributions
mean = 50
std_dev = 10
prob = norm.cdf(60, mean, std_dev) # Probability of a value ≤ 60
print(f"Probability of a value ≤ 60: {prob:.4f}")
Summary:
✓
⬛ Understanding data types and their significance in ML.
✓
⬛ Learning sampling techniques for unbiased data collection.
✓
⬛ Applying descriptive statistics to summarize data distributions.
✓
⬛ Gaining foundational knowledge in probability for ML models.
Page | 34
ACTIVITY LOG FOR THE FORTH WEEK
Day Person
Brief description of the daily Learning Outcome In-Charge
activity Signature
Page | 35
WEEKLY REPORT
WEEK–4 (From Dt………..…..to Dt................. )
Detailed Report:
Page | 36
Advanced Statistics for Machine Learning
This week covers advanced statistical concepts used in Machine Learning (ML) to make data-
driven decisions and evaluate models effectively.
1. Marginal Probability
2. Joint Probability
3. Conditional Probability
4. Bayes' Theorem
Page | 37
Inferential Statistics
Inferential statistics help make predictions about a population based on a sample.
1. Point Estimation
Definition: Estimating a population parameter using a single value (e.g., sample mean for
population mean).
Example: Predicting the average salary of all data scientists using a small sample.
2. Interval Estimate
Application in ML
Hypothesis Testing
Used to validate claims or compare models.
Page | 38
Test Type Usage Example
Comparing means of two groups (e.g., average test scores before &
T-Test
after ML training).
Used for categorical data analysis (e.g., checking correlation between
Chi-Square Test
gender and product preference).
ANOVA (Analysis of Used to compare more than two groups (e.g., performance of multiple
Variance) ML models).
Example in Python:
1. Understanding Entropy
X, y = load_iris(return_X_y=True)
tree = DecisionTreeClassifier(criterion='entropy')
tree.fit(X, y)
Page | 39
Confusion Matrix
Used to evaluate classification models.
1. Accuracy:
Accuracy=(TP+TN)/(TP+TN+FP+FN)
Precision=TP/(TP+FP)
Recall=TP/(TP+FN)
F1=(2×Precision×Recall)/(Precision+Recall)
3. Example in Python
from sklearn.metrics import confusion_matrix, classification_report
import numpy as np
cm = confusion_matrix(y_true, y_pred)
print("Confusion Matrix:\n", cm)
print("\nClassification Report:\n", classification_report(y_true, y_pred))
Page | 40
Summary
✓
⬛ Marginal & Joint Probability, Bayes’ Theorem for probabilistic ML models.
✓
⬛ Inferential Statistics for making data-driven predictions.
✓
⬛ Hypothesis Testing for validating ML model improvements.
✓
⬛ Entropy & Information Gain for decision trees.
✓
⬛ Confusion Matrix & Performance Metrics for evaluating classification models.
Page | 41
ACTIVITY LOG FOR THE FIFTH WEEK
Day Person
Brief description of the daily Learning Outcome In-Charge
activity Signature
Page | 42
WEEKLY REPORT
WEEK–5(From Dt………..…..to Dt.................. )
Detailed Report:
Page | 43
Linear Regression
Introduction and Working of Linear Regression
Linear Regression is a fundamental supervised learning algorithm used for predicting continuous
values. It establishes a relationship between a dependent variable Y and one or more independent
variables X. The relationship is modeled as:
Y=mX+c+ϵ
where:
1. Simple Linear Regression: A single independent variable is used to predict the dependent
variable.
2. Multiple Linear Regression: Multiple independent variables are used for prediction.
Linear regression can be applied to analyze Titanic survival predictions by examining variables like
passenger class, age, fare, etc., to estimate the likelihood of survival.
Logistic Regression
Understanding Logistic Regression Curve
Logistic Regression is a supervised learning algorithm used for binary classification problems
(e.g., spam detection, fraud detection). Unlike linear regression, it predicts probabilities rather than
continuous values.
P(Y=1∣X)=1/(1+e-(β0+β1X))
Page | 44
where:
σ(z)=1 /(1+e-z)
z is a linear combination of the input features: z=w0+w1x1+w2x2+⋯+wnxn (where w are the
weights and x are the input features)
σ(z) maps any real number to a value between 0 and 1.
S-shaped curve (also called the "logistic curve").
Output is always between 0 and 1 (suitable for probabilities).
If z=0 the probability is 0.5 (neutral decision boundary).
As z→∞, the probability approaches 1.
As z→−∞ the probability approaches 0.
Hands-on Demo
Page | 45
Select Relevant Features
Age (numerical)
Sex (categorical: male/female)
Pclass (categorical: 1st, 2nd, 3rd class)
Since Age has different scales compared to other features, we apply Standardization.
scaler = StandardScaler()
X_train = scaler.fit_transform(X_train)
X_test = scaler.transform(X_test)
Page | 46
print(f'Accuracy: {accuracy:.2f}')
# Confusion Matrix
conf_matrix = confusion_matrix(y_test, y_pred)
print("Confusion Matrix:\n", conf_matrix)
# Classification Report
print("Classification Report:\n", classification_report(y_test, y_pred))
Results Interpretation
Decision Trees
What is Classification?
Classification is a supervised learning task where the goal is to categorize data points into
predefined classes (labels).
A Decision Tree is a tree-like structure where internal nodes represent features, branches represent
decision rules, and leaf nodes represent the outcome.
Key Terminologies:
Page | 47
How Decision Trees Work?
1. Start with the root node containing all training samples.
2. Choose the best feature to split the data based on a splitting criterion (e.g., Gini Index,
Entropy).
3. Recursively split data into branches until all data points belong to a specific class.
4. Prune the tree to improve generalization.
Decision trees use different criteria to decide the best feature to split on. Two popular methods are
Entropy and Gini Index.
where pip_ipi is the probability of class iii. A lower entropy value indicates a purer node.
Information Gain (IG) is used to determine the best feature for splitting:
where Sv are the subsets after splitting. A feature with high information gain is preferred.
Gini Index
Gini=1−∑pi 2
Page | 48
ACTIVITY LOG FOR THE SIXTH WEEK
Day Person
Brief description of the daily Learning Outcome In-Charge
activity Signature
Page | 49
WEEKLY REPORT
WEEK–6(From Dt……….…..to Dt................... )
Detailed Report:
Page | 50
Random Forest
Introduction to Random Forest and Its Advantages
Random Forest is an ensemble learning method that builds multiple decision trees and combines
their outputs to improve accuracy and reduce over fitting.
✓
⬛ Reduces Overfitting: Multiple trees generalize better.
✓
⬛ Handles Missing Values: Works well even with incomplete data.
✓
⬛ Works on Large Datasets: Efficient in handling high-dimensional data.
✓
⬛ Handles Non-linearity: Unlike simple decision trees, it captures complex relationships.
Page | 51
Choosing the Best k Value:
Naive Bayes
Naive Bayes Working & Formula
Naive Bayes is a probabilistic classifier based on Bayes’ Theorem:
P(A∣B) = (P(B∣A)P(A))/P(B)
where:
The "Naive" assumption is that features are independent, which simplifies calculations.
Page | 52
The goal is to classify whether a patient has diabetes using Naive Bayes.
Features include glucose levels, blood pressure, insulin, age, etc.
Why Naive Bayes? It works well even with small datasets and is computationally efficient.
Mathematical Representation:
wX+b=0
y=sign(wX+b)
where:
w = Weight vector
X = Feature vector
b = Bias
Popular Kernels:
Page | 53
1. Linear Kernel: For linearly separable data.
2. Polynomial Kernel: For moderately complex data.
3. Radial Basis Function (RBF) Kernel: Commonly used for highly non-linear problems.
Page | 54
ACTIVITY LOG FOR THE SEVENTH WEEK
Day Person
Brief description of the daily Learning Outcome In-Charge
activity Signature
Page | 55
WEEKLY REPORT
WEEK–7(From Dt………..…..to Dt.................. )
Detailed Report:
Page | 56
Unsupervised Learning & Clustering
In this week, we explore unsupervised learning techniques, particularly clustering and
association rule mining. These methods help find hidden patterns and relationships in data
without labeled outputs.
Introduction to Clustering
What is Clustering?
Clustering is an unsupervised learning technique used to group similar data points into clusters
based on their features.
1. K-Means Clustering
o Partition-based clustering method.
o Assigns data points to K clusters based on similarity.
2. Hierarchical Clustering
o Builds a hierarchy of clusters.
o Can be Agglomerative (bottom-up) or Divisive (top-down).
3. DBSCAN (Density-Based Clustering)
o Groups points based on density.
o Good for irregularly shaped clusters.
4. Gaussian Mixture Model (GMM)
o Uses probability distributions for clustering.
o More flexible than K-Means.
K-Means Clustering
Understanding the K-Means Algorithm
⬛
✓ Fast and Scalable for large datasets.
✓
⬛ Works well with clear, well-separated clusters.
+ Sensitive to initial centroid selection.
+ Fails on non-spherical or imbalanced clusters.
Hierarchical Clustering
Types of Hierarchical Clustering
1. Agglomerative (Bottom-Up)
o Start with individual points and merge closest clusters.
2. Divisive (Top-Down)
o Start with all points in one cluster and split recursively.
Page | 58
Steps:
1. Compute distance matrix & apply Agglomerative clustering.
2. Plot the dendrogram to find optimal clusters.
3. Assign cluster labels & visualize the results.
✓
⬛ Doesn’t require specifying K in advance.
✓
⬛ Produces a tree-like structure for better interpretability.
+ Computationally expensive for large datasets.
+ Sensitive to noisy data & outliers.
Apriori Algorithm
The Apriori Algorithm is used to find frequent item sets in transactional datasets using:
Lift(A→B)=Confidence(A→B) / Support(B)
Page | 59
Consider a database of transactions in a supermarket:
T1 Milk, Bread
Page | 60
(3/5)/(4/5)=75%
Since 75% confidence meets the minimum threshold of 70%, the rule is kept.
Rule: {Diaper} ⇒ {Beer}: The confidence is:
Confidence(Diaper⇒Beer)=Support(Diaper,Beer)/Support(Diaper)
=(3/5)/(4/5)=75%
This rule is also kept because it meets the confidence threshold.
Example of Apriori Results
1. Frequent Itemsets:
o Single items: {Milk}, {Bread}, {Diaper}, {Beer}
o Pairs: {Milk, Bread}, {Milk, Diaper}, {Milk, Beer}, {Bread, Diaper}, {Bread,
Beer}, {Diaper, Beer}
2. Association Rules:
o {Milk} ⇒ {Bread} (75% confidence)
o {Diaper} ⇒ {Beer} (75% confidence)
Conclusion
The Apriori algorithm is widely used in market basket analysis and other applications involving
association rule mining. By identifying frequent itemsets and deriving strong association rules,
businesses can understand customer purchasing behavior and make informed decisions regarding
product placements, promotions, and recommendations.
Summary:
✓
⬛ Introduction to Clustering → Types & real-world applications.
✓
⬛ K-Means Clustering → Working, pros/cons & customer segmentation demo.
✓
⬛ Hierarchical Clustering → Agglomerative vs. Divisive methods & practical example.
✓
⬛ Association Rule Mining → Apriori Algorithm for market basket analysis.
Page | 61
ACTIVITY LOG FOR THE EIGTH WEEK
Day Person
Brief description of the daily Learning Outcome In-Charge
activity Signature
Page | 62
WEEKLY REPORT
WEEK–8(From Dt………..…..to Dt:Dt....................)
Detailed Report:
Page | 63
Reinforcement Learning (RL)
Reinforcement Learning (RL) is a type of machine learning where an agent learns to make
decisions by interacting with an environment and receiving rewards or penalties based on its
actions. This week covers the fundamentals of RL, including Markov Decision Processes (MDP),
Q-Learning, the Bellman Equation, and hands-on projects.
A Markov Decision Process (MDP) provides a mathematical framework for modeling decision-
making in sequential environments where outcomes are partly random and partly controlled by an
agent’s actions.
MDP Components:
✓
⬛ Defines an environment mathematically, helping in designing RL algorithms.
✓
⬛ Helps the agent learn optimal decision-making strategies over time.
✓
⬛ Forms the basis for Q-Learning, Deep Q-Networks (DQN), and Policy Gradient Methods.
Q(s,a)=Q(s,a)+α[R+γmaxQ(s′,a′)−Q(s,a)]
where:
The Bellman Equation describes the optimal value function for an MDP, helping in determining
the best possible action for an agent.
V∗(s)=maxa[R(s,a)+γs′∑P(s′∣s,a)V∗(s′)]
where:
✓
⬛ Helps compute the optimal policy for decision-making.
✓
⬛ Forms the foundation of Q-Learning and Deep Q-Networks (DQN).
✓
⬛ Used to evaluate expected future rewards.
Implementing Q-Learning
Implementing Q-Learning in a Simulated Environment
Page | 65
Environment: OpenAI Gym (e.g., FrozenLake, Taxi-v3).
Algorithm: Use Q-Learning to teach an agent to navigate the environment.
c)
’ Project 1: Training an RL Agent to Play a Self-Play Game
c)
’ Project 2: Self-Driving Car Simulation
Page | 66
Summary:
✓
⬛ Markov Decision Process (MDP) → Foundation of RL, Q-learning intro.
✓
⬛ The Bellman Equation → Understanding and applying the Bellman equation.
✓
⬛ Implementing Q-Learning → Train an RL agent, Counter-Strike example.
✓
⬛ RL Projects → Apply Q-Learning to self-play games & real-world scenarios.
Page | 67
ACTIVITY LOG FOR THE NINETH WEEK
Day Person
Brief description of the daily Learning Outcome In-Charge
activity Signature
Page | 68
WEEKLY REPORT
WEEK–9(From Dt………..…..to Dt...................)
Detailed Report:
Page | 69
Machine Learning Project Development (Part 1)
This week focuses on the end-to-end process of developing a machine learning (ML) project,
from data preprocessing to model deployment. By the end of this module, you will be able to
prepare data, train & evaluate models, and deploy them to the cloud or a web interface.
Data Preprocessing
Before training a machine learning model, the data needs to be cleaned and transformed into a
format suitable for the model. This step significantly impacts the accuracy and performance of
the model.
1. Data Cleaning
c)
’ Techniques to Handle Missing Values
’c Handling Outliers
)
c
)’Why Encode Categorical Variables?
Machine learning models work with numerical values. Categorical data must be converted into
numerical format.
c Types of Encoding
’
)
Label Encoding (Converts categories into numerical values, e.g., Male → 0, Female → 1).
Page | 70
One-Hot Encoding (Creates a binary column for each category).
Ordinal Encoding (For ordered categories like Low, Medium, High → 1, 2, 3).
’ hy Scale Features?
)
W
Models like SVM, k-NN, and Gradient Descent-based algorithms are affected by varying
scales.
Ensures that features contribute equally to the learning process.
)
’
c Scaling Techniques
c)
’ Python Implementation Example
# Load dataset
df = pd.read_csv('data.csv')
print(df.head())
The dataset is split to train the model and evaluate its performance.
Train Set (70-80%) → Used to train the model.
Test Set (20-30%) → Used to check model performance.
Validation Set (Optional, 10-15%) → Fine-tunes hyperparameters before testing.
Page | 71
’
)
c Splitting Data in Python
# Splitting dataset
X = df.drop('Target', axis=1)
y = df['Target']
2. Model Training
’
)
c Choosing the Right Model
c)
’ Training a Sample Model (Random Forest)
# Train model
model = RandomForestClassifier(n_estimators=100, random_state=42)
model.fit(X_train, y_train)
’c Classification Metrics
)
’c Regression Metrics
)
)
c’ Evaluation Example in Python
# Predictions
Page | 72
y_pred = model.predict(X_test)
# Evaluation
accuracy = accuracy_score(y_test, y_pred)
print("Accuracy:", accuracy)
print(classification_report(y_test, y_pred))
Makes the model available for real-world applications via an API or web interface.
Can be hosted on cloud platforms like AWS, Azure, or Google Cloud.
’c Deployment Platforms
)
c)
’ Basic Model Deployment with Flask
@app.route('/predict', methods=['POST'])
def predict():
data = request.get_json()
prediction = model.predict([data['features']])
return jsonify({'prediction': prediction.tolist()})
import streamlit as st
import pickle
# Load model
model = pickle.load(open('model.pkl', 'rb'))
Page | 73
# UI
st.title("ML Model Deployment")
input_data = st.text_input("Enter Features:")
if st.button("Predict"):
prediction = model.predict([eval(input_data)])
st.write("Prediction:", prediction)
’
c
) Run the app:
Summary:
✓
⬛ Data Preprocessing → Cleaning, handling missing values, encoding & scaling.
✓
⬛ Model Training & Testing → Splitting data, training, and evaluating models.
✓
⬛ Model Deployment → Deploying models on the cloud & building a web interface.
Page | 74
ACTIVITY LOG FOR THE TENTH WEEK
Day Person
Brief description of the daily Learning Outcome In-Charge
activity Signature
Page | 75
WEEKLY REPORT
WEEK–10(From Dt………..…..to Dt.................. )
Detailed Report:
Page | 76
Python Basics: A Comprehensive Guide
Python is a high-level, interpreted programming language known for its simplicity and
readability. It is widely used in web development, data science, artificial intelligence, machine
learning, automation, and scripting.
c)
’ Python Installation & Running Python Code
c)
’ Variable Declaration Example
x = 10
name = "Alice"
pi = 3.14
Page | 77
Rules for Variable Names (Identifiers)
✓
P Must start with a letter (A-Z or a-z) or an underscore _.
✓
P Can contain letters, digits (0-9), and underscores.
✓
P Cannot be a reserved keyword (e.g., if, else, while).
✓
P Case-sensitive (Age and age are different).
)
’c Checking Data Type of a Variable
x = 5
print(type(x)) # Output: <class 'int'>
c)
’ Example: Taking Integer Input
c)
’ Using Format Strings (f-strings)
name = "Alice"
age = 25
print(f"My name is {name} and I am {age} years old.")
Page | 78
4. Operators in Python
Types of Operators
c Example
’
)
x = 10
y = 3
print(x + y) # 13
print(x ** y) # 1000 (10^3)
c Example
’
)
x = 5
y = 10
print(x > 2 and y < 15) # True
Page | 79
5. Conditional Statements
1. if Statement
’
c
) Basic if Condition
age = 18
if age >= 18:
print("You are eligible to vote!")
2. if-else Statement
3. if-elif-else Statement
)
c’ Example with Multiple Conditions
if operation == '+':
print("Result:", a + b)
elif operation == '-':
print("Result:", a - b)
elif operation == '*':
print("Result:", a * b)
elif operation == '/':
print("Result:", a / b)
Page | 80
else:
print("Invalid operation")
Conclusion
’● Python is beginner-friendly and powerful.
◉
●’◉ Understanding variables, operators, and conditional statements is fundamental.
’ Practice writing small programs to reinforce learning.
◉
●
Page | 81
ACTIVITY LOG FOR THE ELEVENTH WEEK
Day Person
Brief description of the daily Learning Outcome In-Charge
activity Signature
Page | 82
WEEKLY REPORT
WEEK–11(From Dt………..…..to Dt.................. )
Detailed Report:
Page | 83
Python: In-depth Explanation with Examples
a = 5 # int
b = 2.5 # float
c = 3 + 4j # complex
c Example
’
)
s = "Python" # String
l = [1, 2, 3] # List
t = (4, 5, 6) # Tuple
Page | 84
2. Lists, Tuples, Sets, and Dictionaries
A. Lists in Python
◆ List Operations
numbers = [1, 2, 3]
print(numbers[1]) # Accessing elements
print(len(numbers)) # Finding length
print(numbers[::-1]) # Reversing a list
B. Tuples in Python
c)
’ Tuples are immutable, ordered collections.
C. Sets in Python
c)
’ Sets are unordered, unique collections.
◆ Set Operations
A = {1, 2, 3}
B = {3, 4, 5}
print(A | B) # Union {1, 2, 3, 4, 5}
print(A & B) # Intersection {3}
print(A - B) # Difference {1, 2}
D. Dictionaries in Python
Page | 85
c)
’ Dictionaries store key-value pairs.
◆ Dictionary Methods
3. Functions in Python
A. Defining Functions
c)
’ Functions help to reuse code efficiently.
def greet(name):
return f"Hello, {name}!"
Page | 86
C. Appending to a File
with open("test.txt", "a") as file:
file.write("\nThis is a new line!")
5. Comprehensions in Python
A. List Comprehension
c)
’ Creating a list with a single line of code.
B. Dictionary Comprehension
squares_dict = {x: x**2 for x in range(5)}
print(squares_dict) # {0: 0, 1: 1, 2: 4, 3: 9, 4: 16}
C. Set Comprehension
unique_numbers = {x for x in [1, 2, 2, 3, 4, 4, 5]}
print(unique_numbers) # {1, 2, 3, 4, 5}
Page | 87
merged = {**dict1, **dict2}
print(merged)
Conclusion
’●
◉ Python provides flexible data types (Lists, Tuples, Sets, and Dicts).
’●◉ Functions improve code reusability.
’ File handling is essential for working with external data.
◉
●
’●
◉ Comprehensions simplify list, dict, and set creation.
Page | 88
ACTIVITY LOG FOR THE TWELVETH WEEK
Day Person
Brief description of the daily Learning Outcome In-Charge
activity Signature
Students works on
Explain programs regarding oops
programs regarding
concepts (inheritance,
Day–3 oops concepts
polymorphism)
(inheritance,
polymorphism)
Page | 89
WEEKLY REPORT
WEEK–12(From Dt………..…..to Dt.................. )
Detailed Report:
Explain packages
Page | 90
Object-Oriented Programming (OOP) in Python
Object-Oriented Programming (OOP) is a programming paradigm that uses objects and classes to
structure code. It helps in organizing code into reusable and scalable components.
’
)cA class is a blueprint for creating objects.
)
’c An object is an instance of a class.
# Defining a Class
class Car:
def init (self, brand, model): # Constructor
self.brand = brand
self.model = model
def show_info(self):
return f"Car: {self.brand}, Model: {self.model}"
# Creating Objects
car1 = Car("Toyota", "Corolla")
car2 = Car("Honda", "Civic")
c)’ Encapsulation means restricting direct access to data and allowing modifications via
methods.
class BankAccount:
def init (self, balance):
self. balance = balance # Private variable
def get_balance(self):
return self. balance
# Creating Object
account = BankAccount(1000)
Page | 91
account.deposit(500)
print(account.get_balance()) # Output: 1500
B. Inheritance
# Parent Class
class Animal:
def sound(self):
return "Animal makes sound"
# Child Class
class Dog(Animal):
def sound(self): # Method Overriding
return "Dog barks"
dog = Dog()
print(dog.sound()) # Output: Dog barks
C. Polymorphism
class Cat:
def sound(self):
return "Meow"
class Dog:
def sound(self):
return "Bark"
# Using Polymorphism
for animal in (Cat(), Dog()):
print(animal.sound())
◆ Both Cat and Dog have a sound() method, but their implementations are different.
D. Abstraction
c)’ Abstraction hides implementation details and only exposes the necessary parts.
class Vehicle(ABC):
@abstractmethod
def start(self):
Page | 92
pass
class Car(Vehicle):
def start(self):
return "Car is starting"
car = Car()
print(car.start()) # Output: Car is starting
class Child(Parent):
def display(self):
return "This is the Child class"
obj = Child()
print(obj.show()) # Parent class method
print(obj.display()) # Child class method
Program: Polymorphism
class Shape:
def area(self):
return 0
class Square(Shape):
def init (self, side):
self.side = side
def area(self):
return self.side * self.side
sq = Square(5) print(sq.area())
# Output: 25
Program: Abstraction
from abc import ABC, abstractmethod
class Animal(ABC):
@abstractmethod
def make_sound(self):
pass
Page | 93
class Dog(Animal):
def make_sound(self):
return "Bark"
dog = Dog()
print(dog.make_sound()) # Output: Bark
Program: Encapsulation
class Student:
def init (self, name, age):
self. name = name # Private attribute
self. age = age # Private attribute
def get_details(self):
return f"Student: {self. name}, Age: {self. age}"
4. Packages in Python
)’
cA package is a collection of Python modules.
c)’ It helps in organizing large projects.
Creating a Package
# math_operations.py
def add(a, b):
return a + b
result = math_operations.add(5, 3)
print(result) # Output: 8
◆ The init .py file tells Python that this directory is a package.
Page | 94
5. Arrays in Python
c)’ Arrays store multiple values of the same type.
)’c Python provides array module for handling arrays.
import array
# Creating an array
arr = array.array('i', [10, 20, 30])
# Accessing elements
print(arr[0]) # Output: 10
print(arr[1]) # Output: 20
# Adding elements
arr.append(40)
# Removing elements
arr.remove(20)
6. Programs on Arrays
A. Find Maximum in an Array
import array
B. Reverse an Array
import array
Page | 95
import array
Conclusion
✓
⬛ OOP concepts help in structuring large programs efficiently.
✓
⬛ Encapsulation, Inheritance, Polymorphism, and Abstraction enhance code reusability and
security.
✓
⬛ Packages allow modular programming and easy maintenance.
✓
⬛ Arrays store multiple values efficiently and support basic operations.
Page | 96
ACTIVITY LOG FOR THE THIRTEENTH WEEK
Day Person
Brief description of the daily Learning Outcome In-Charge
activity Signature
Page | 97
WEEKLY REPORT
WEEK–13(From Dt………..…..to Dt.................. )
Detailed Report:
Explain about NumPy for numerical computing
Page | 98
Python for Data Science: NumPy, Pandas,
Matplotlib, and Seaborn
Python is widely used for data science and numerical computing. The key libraries for these tasks
include NumPy, Pandas, Matplotlib, and Seaborn. This guide covers these libraries in-depth,
along with practical Python scripts and Exploratory Data Analysis (EDA) on the Titanic dataset.
✓
⬛ Efficient handling of large datasets
✓
⬛ Vectorized operations (faster than Python lists)
✓
⬛ Supports mathematical and statistical operations
Installing NumPy
pip install numpy
# 1D Array
arr1 = np.array([1, 2, 3, 4, 5])
print(arr1) # Output: [1 2 3 4 5]
# 2D Array (Matrix)
arr2 = np.array([[1, 2, 3], [4, 5, 6]])
print(arr2)
NumPy Operations
# Arithmetic Operations
a = np.array([10, 20, 30])
b = np.array([1, 2, 3])
print(a + b) # Output: [11 22 33]
Page | 99
Statistical Functions
data = np.array([1, 2, 3, 4, 5])
print(np.mean(data)) # Mean
print(np.median(data)) # Median
print(np.std(data)) # Standard Deviation
Installing Pandas
pip install pandas
df = pd.DataFrame(data)
print(df)
Page | 100
# Display data types
print(df.dtypes)
# Summary Statistics
print(df.describe())
Installing Matplotlib
pip install matplotlib
Basic Plotting
import matplotlib.pyplot as plt
Bar Chart
categories = ["A", "B", "C"]
values = [30, 50, 80]
Page | 101
Installing Seaborn
pip install seaborn
Creating a Histogram
import seaborn as sns
# Sample Data
data = [10, 20, 20, 30, 30, 30, 40, 50, 60]
# Histogram
sns.histplot(data, bins=5, kde=True)
plt.title("Histogram Example")
plt.show()
print("Mean:", np.mean(data))
print("Median:", np.median(data))
df = pd.read_csv("data.csv")
Page | 102
5. Performing EDA on Titanic Dataset
Step 1: Load the Dataset
import pandas as pd
# Countplot of Survival
sns.countplot(x="Survived", data=df)
plt.title("Survival Count")
plt.show()
Page | 103
Step 6: Train a Simple Model
from sklearn.model_selection import train_test_split
from sklearn.ensemble import RandomForestClassifier
from sklearn.metrics import accuracy_score
# Train-test split
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2,
random_state=42)
# Train model
model = RandomForestClassifier()
model.fit(X_train, y_train)
# Predictions
y_pred = model.predict(X_test)
# Accuracy
print("Model Accuracy:", accuracy_score(y_test, y_pred))
Conclusion
✓
⬛ NumPy for numerical computing
✓
⬛ Pandas for data manipulation
✓
⬛ Matplotlib & Seaborn for data visualization
⬛
✓ EDA on Titanic Dataset c÷
X
Page | 104
ACTIVITY LOG FOR THE FOURTEENTH WEEK
Day Person
Brief description of the daily Learning Outcome In-Charge
activity Signature
Students Implement
Day-2 Implement Real-World ML Real-World ML Project
Project (capstone project)
Students Implement
Day–3 Implement Optimization & Optimization &
Hyperparameter Tuning Hyperparameter Tuning
Day–5 Students
Implement Real-World ML Project
Implement Real-
(capstone project) (continued...)
World ML Project
Day–6 Students
Implement Real-World ML Project
Implement Real-
(capstone project) (continued...)
World ML Project
Page | 105
WEEKLY REPORT
WEEK–14(From Dt………..…..to Dt.................. )
Detailed Report:
Page | 106
Advanced Machine Learning Algorithms &
Capstone Project Implementation
1. Advanced Machine Learning Algorithms
Once you understand basic supervised and unsupervised learning algorithms, you can explore more
advanced ML techniques to improve model performance and solve complex real-world problems.
✓
⬛ Bagging (Bootstrap Aggregating)
✓
⬛ Boosting (Adaptive Learning)
✓
⬛ Stacking (Meta-Learning)
c)’ Random Forest is an extension of Decision Trees where multiple trees are built using different
subsets of data.
’)c The final prediction is made by majority voting (for classification) or averaging (for
regression).
Implementation:
# Load dataset
data = load_iris()
X, y = data.data, data.target
# Split dataset
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2,
random_state=42)
Page | 107
# Predictions
y_pred = model.predict(X_test)
# Accuracy
print("Random Forest Accuracy:", accuracy_score(y_test, y_pred))
)’c Boosting is an iterative technique that adjusts model weights to focus on hard-to-classify
cases.
c) ’ Gradient Boosting & XGBoost are popular implementations.
# Predictions
y_pred = xgb_model.predict(X_test)
# Accuracy
print("XGBoost Accuracy:", accuracy_score(y_test, y_pred))
PCA Example:
# Apply PCA
pca = PCA(n_components=2)
X_pca = pca.fit_transform(X)
Page | 108
from sklearn.cluster import KMeans
TF-IDF Example:
1. Filter Methods
Page | 109
2. Wrapper Methods
3. Embedded Methods
4. Hybrid Methods
Filter methods use statistical techniques to evaluate feature importance before training the
model.
◆ Common Techniques:
c)
’ Example (Using Scikit-Learn in Python)
Wrapper methods train models iteratively using different feature subsets and select the best-
performing subset.
◆ Common Techniques:
Forward Selection: Starts with no features, adds the best feature iteratively.
Backward Elimination: Starts with all features, removes least significant one at each step.
Recursive Feature Elimination (RFE): Eliminates least important features recursively.
)
c’ Example (Using RFE in Python)
Page | 110
model = RandomForestClassifier()
rfe = RFE(model, n_features_to_select=5) # Selects top 5 features
X_new = rfe.fit_transform(X, y)
Embedded methods integrate feature selection into the model training process.
◆ Common Techniques:
Lasso Regression (L1 Regularization): Shrinks less important feature coefficients to zero.
Decision Tree Feature Importance: Trees naturally rank features based on splits.
c)
’ Example (Using Lasso in Python)
c)
’ Example (Using Feature Importance in Random Forest)
Hybrid methods combine filter, wrapper, and embedded methods to optimize feature selection.
◆ Example Workflow:
Page | 111
4. Feature Selection vs Feature Extraction
Feature Selection Feature Extraction
Selects a subset of existing features Creates new features from existing ones
Does not alter original features Transforms features (e.g., PCA, Autoencoders)
Example: Removing redundant columns Example: Reducing dimensions using PCA
6. Conclusion
Feature selection is crucial for improving model performance.
Different methods (Filter, Wrapper, Embedded) have trade-offs.
Hybrid methods offer the best results but are computationally expensive.
)’c Predict house prices based on features such as area, number of bedrooms, location, etc.
c) ’ Use regression techniques to develop a predictive model.
Page | 112
Model Evaluation: Check performance metrics
Deployment: Deploy the model using Flask/Streamlit
c)
’ Load house price dataset and clean it.
import pandas as pd
# Load dataset
df = pd.read_csv("house_prices.csv")
c)
’ Visualizing correlations in data
# Correlation Heatmap
plt.figure(figsize=(10,6))
sns.heatmap(df.corr(), annot=True, cmap="coolwarm")
plt.show()
c)
’ Train a Linear Regression Model
# Train-test split
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2,
random_state=42)
# Train Model
model = LinearRegression()
model.fit(X_train, y_train)
Page | 113
# Predictions
y_pred = model.predict(X_test)
# Evaluation
print("MAE:", mean_absolute_error(y_test, y_pred))
print("R2 Score:", r2_score(y_test, y_pred))
c)
’ Deploy the trained model using Flask
@app.route("/predict", methods=["POST"])
def predict():
data = request.json
prediction = model.predict([[data["area"], data["bedrooms"],
data["bathrooms"],
data["location_New York"], data["location_Los
Angeles"]]])
return jsonify({"Predicted Price": prediction[0]})
3. Conclusion
✓
⬛ Covered Advanced ML Algorithms (Ensemble Learning, Clustering, NLP, etc.)
✓
⬛ Built a Real-World House Price Prediction Model
✓
⬛ Performed Data Preprocessing, Model Training, and Deployment
Page | 114
ACTIVITY LOG FOR THE FIFTEENTH WEEK
ML Engineer Interview
Day–4 Preparation Students learns about
interview questions
Page | 115
WEEKLY REPORT
WEEK–15(From Dt………..…..to Dt.................. )
Detailed Report:
Page | 116
Machine Learning Engineer: Role,
Responsibilities & Career Guide
1. Machine Learning Engineer Role & Responsibilities
A Machine Learning Engineer (MLE) is responsible for designing, building, and deploying
ML models to solve real-world problems. They work at the intersection of software
engineering, data science, and artificial intelligence.
✓
⬛ Data Collection & Preprocessing
✓
⬛ Model Development & Training
✓
⬛ Model Evaluation & Validation
✓
⬛ Model Deployment & Monitoring
✓
⬛ Collaboration & Documentation
'•/¸
.
˙½7 Machine Learning Engineer with 3+ years of experience in developing and deploying
scalable ML models. Expertise in deep learning, NLP, and cloud-based ML solutions. Passionate
about solving business problems using AI-driven technologies.
✓
⬛ Machine Learning Basics – Supervised vs. Unsupervised Learning, Overfitting, Bias-
Variance Tradeoff
✓
⬛ Deep Learning – CNNs, RNNs, Transformers, Attention Mechanisms
✓
⬛ Mathematics & Statistics – Probability, Linear Algebra, Gradient Descent
✓
⬛ Feature Engineering – Handling missing values, categorical encoding, dimensionality
reduction
✓
⬛ Model Evaluation – Precision, Recall, AUC-ROC, RMSE, Log Loss
✓
⬛ System Design – Building ML pipelines, scalable architectures
✓
⬛ Deployment & MLOps – Docker, Kubernetes, Model Monitoring
Example Question:
¸
/'
.
˙•7½ Implement a function that normalizes a given dataset using Min-Max Scaling
Page | 118
import numpy as np
def min_max_scaling(data):
return (data - np.min(data)) / (np.max(data) - np.min(data))
# Example Usage
data = np.array([10, 20, 30, 40, 50])
scaled_data = min_max_scaling(data)
print(scaled_data)
✓
⬛ Practice ML coding questions on Leetcode (ML Section), HackerRank, Kaggle Notebooks
⬛
✓ ML Projects with GitHub Repositories
✓
⬛ Jupyter Notebooks & Kaggle Datasets
✓
⬛ Blogs on ML topics (Medium, Dev.to, Hashnode)
✓
⬛ Open-source contributions to ML libraries (Scikit-Learn, TensorFlow, Hugging Face)
□
˙
н House-Price-Prediction/
├── □
н̇ data/ # Raw & processed datasets
├── □
˙ notebooks/
н # Jupyter Notebooks
├── ˙
□
н models/ # Trained model files
˙
□ scripts/
├── н # Python scripts for training & inference
├── requirements.txt # Dependencies (pandas, sklearn, flask)
├── README.md # Project Overview & Instructions
’)c Tip: Keep your README well-documented with model results, graphs, and deployment
steps.
Page | 119
5. Summary & Next Steps
˙•
¸'./
7 ½ Role of an ML Engineer – Model building, deployment, and optimization
.̧
' /½˙•7 How to Build an ML Resume – Highlight skills, projects, and experience
.̧
' /½˙•7 ML Interview Preparation – Coding practice, ML theory, system design
¸ /'
.
˙•7½ Building an Online Portfolio – GitHub, Kaggle, Blogs, Open-source contributions
Page | 120
CHAPTER5: OUTCOMES DESCRIPTION
At Epro academy and Glossary Softech all staff members and Management are supportive as
an Intern My role is to attend on time and maintain self-discipline and dedication towards of
Training program.
Epro academy and Glossary soft tech is best EdTech Company having Infrastructure for both
Online and offline Trainings. I have done good Projects with my co-Interns and Staff. I have
gone through the wonderful training experience.
I have learned about machine learning topics with real time examples, python related to
machine learning curriculum and so many things.
I have done live hands-on Particles and implemented
examples of machine learning techniques I have made one real time project of House Price
Prediction Model using Linear Regression.
At Epro academy and Glossary Softech I have learned about team management and leadership
skills in order to complete the project.
Coordinating with my team and assigning roles for them is very challenging
task but with the support of my mentor and Team leader it will become easy to successfully
manage my team in order to reach our common goal.
Page | 121
Describe how you could improve your communication skills
At Epro academy and Glossary Softech I have improved my oral and written communication
skills by involving in real time projects and maintaining coordination with the team members it
improves my confidence level.at Epro while after taking suggestions from the seniors iam able
to control my anxiety.
The main thing is understanding others, getting understood by others, extempore speech, ability
to articulate the key points, closing the conversation, maintaining niceties and protocols,
greeting, thanking and appreciating others, etc are I have learned at Epro academy. hence, I
have improvised myself in all aspects as an intern.)
Describe how could you enhance your abilities in group discussions, participation in teams,
contribution as a team member, leading a team/activity.
As an Intern In group discussion, I have learned how to Control voice and tone while discussing
with my team members and how to respect others’ opinions and thoughts about the Topic we are
discussing.
I have learned about how respect other team members Ideas and way of
convincing them on a right path. while doing project work, I have learned how to lead the team
from my faculty members.
Describe the technological developments you have observed and relevant to the subject area
of training (focus on digital technologies relevant to your job role)
While doing project work as an intern most of the things are done coding, In
Machine learning I have also learned about latest trends in software field job related knowledge.
I have learned about latest codes and lines of programs which makes very easy to made train a
model.
Page | 122
Page | 123
Page | 124
Student Self Evaluation of the Short-Term Internship
Date of Evaluation:
1 Oral communication 1 2 3 4 5
2 Written communication 1 2 3 4 5
3 Proactiveness 1 2 3 4 5
4 Interaction ability with community 1 2 3 4 5
5 Positive Attitude 1 2 3 4 5
6 Self-confidence 1 2 3 4 5
7 Ability to learn 1 2 3 4 5
8 Work Plan and organization 1 2 3 4 5
9 Professionalism 1 2 3 4 5
10 Creativity 1 2 3 4 5
11 Quality of work done 1 2 3 4 5
12 Time Management 1 2 3 4 5
13 Understanding the Community 1 2 3 4 5
14 Achievement of Desired Outcomes 1 2 3 4 5
15 OVERALLPERFORMANCE 1 2 3 4 5
Page | 125
Evaluation by the Supervisor of the Intern Organization
Date of Evaluation:
Please note that your evaluation shall be done independent of the Student’s self-
evaluation
1 Oral communication 1 2 3 4 5
2 Written communication 1 2 3 4 5
3 Proactiveness 1 2 3 4 5
4 Interaction ability with community 1 2 3 4 5
5 Positive Attitude 1 2 3 4 5
6 Self-confidence 1 2 3 4 5
7 Ability to learn 1 2 3 4 5
8 Work Plan and organization 1 2 3 4 5
9 Professionalism 1 2 3 4 5
10 Creativity 1 2 3 4 5
11 Quality of work done 1 2 3 4 5
12 Time Management 1 2 3 4 5
13 Understanding the Community 1 2 3 4 5
14 Achievement of Desired Outcomes 1 2 3 4 5
15 OVERALL PERFORMANCE 1 2 3 4 5
Page | 126
Internal & External Evaluation for Semester Internship
Objectives:
Explore career alternatives prior to graduation.
To assess interests and abilities in the field of study.
To develop communication, interpersonal and other critical skills in the future job.
To acquire additional skills required for the world of work.
To acquire employment contacts leading directly to a
full-time job following graduation from college.
Assessment Model:
There shall be both internal evaluation and external evaluation
The Faculty Guide assigned is in-charge of the learning activities of the students
and for the comprehensive and continuous assessment of the students.
The assessment is to be conducted for 200 marks. Internal Evaluation for
50marksandExternalEvaluationfor150marks
The number of credits assigned is 12. Later the marks shall be convertedinto
grades and grade points to include finally in the SGPA and CGPA.
The weightings for Internal Evaluation shall be:
o Activity Log 10marks
o Internship Evaluation 30marks
o Oral Presentation 10marks
The weightings for External Evaluation shall be:
o Internship Evaluation 100marks
o Viva-Voce 50marks
The External Evaluation shall be conducted by an Evaluation Committee
comprising of the Principal, FacultyGuide, Internal Expert and External Expert
nominated by the affiliating University. The Evaluation Committee shall also
consider the grading given by the Supervisor of the Intern Organization.
Activity Log is the record of the day-to-day activities. The Activity Log is
assessed on an individual basis, thus allowing for individual members within
groups to be assessed this way. The assessment will take into consideration
Page | 127
the individual student’s involvement in the assigned work.
While evaluating the student’s Activity Log, the following shall be considered-
a. The individual student’s effort and commitment.
b. Theoriginalityandqualityoftheworkproducedbytheindividualstudent.
c. The student’s integration and co-operation with the work assigned.
d. The completeness of the Activity Log.
The Internship Evaluation shall include the following components and
based on WEEKLY REPORTs and Outcomes Description
a. Description of the Work Environment.
b. RealTime Technical Skills acquired.
c. Managerial Skills acquired.
d. Improvement of Communication Skills.
e. Team Dynamics
f. Technological Developments recorded.
Page | 128
MARKSSTATEMENT
(To be used by the Examiners)
INTERNALASSESSMENTSTATEMENT
1. Activity Log 10
2. Internship Evaluation 30
3. Oral Presentation 10
GRANDTOTAL 50
Page | 129
EXTERNAL ASSESSMENT STATEMENT
Maximum Marks
Sl.No Evaluation Criterion Marks Awarded
1. Internship Evaluation 80
For the grading giving by the Supervisor of the I
2. 20
intern Organization
3. Viva-Voce 50
TOTAL 150
GRANDTOTAL(EXT.50M+INT.100M) 200
Page | 130