Computer Vision LP 2024-25 New-1.1 Final
Computer Vision LP 2024-25 New-1.1 Final
FMTH0301/Rev.5.3
Course Plan
Brief description of the course: This course encompasses principles and techniques to
interpret and analyze visual data, such as images and videos in the field of artificial
intelligence. It's based on the idea of replicating how humans see and understand the world.
2. Analyze principles and techniques of computer vision through features and filtering.
1
School of Computer Science and Engineering
2
School of Computer Science and Engineering
Eg: 1.2.3: Represents Program Outcome ‘1’, Competency ‘2’, and Performance Indicators ‘3’.
3
School of Computer Science and Engineering
Course Content
Content Hrs
Unit – 1
Introduction: Computer Vision Overview, Pixels and image representation, 4hrs
1 Filters:
Linear systems, Convolutions and cross-correlations; Lab: Basics, Filters
Features and filtering: Edge detection: Gaussian, Sobel filters, Canny edge 8hrs
detector, Features and fitting: RANSAC Local features, Harris corner
2
detection, Feature descriptors: Difference of gaussians, Scale invariant
feature transform; Lab: Filters, Edges, Features
Semantic segmentation part I: Perceptual grouping, Agglomerative 3hrs
3
clustering, Super pixels and over segmentation;
Unit – 2
Semantic segmentation part II: Clustering: K-means, Mean shift; Visual Bag 3 hrs
3 of Words: Texture features, Visual bag of words; Lab: Resizing, clustering,
recognition.
Motion: Optical Flow, Lucas-Kanade method, Horn-Schunk Method, 6hrs
4 Pyramids for large motion, Tracking: Feature Tracking, Lucas KanadeTomasi
(KLT) tracker; Lab: Object detection, optical flow.
Advanced Techniques: Image stitching, Image pyramids, Object recognition, 6hrs
5
Dimensionality reduction, Face identification, Detecting objects by parts.
Reference Books:
1. Richard Szeliski, Computer Vision: Algorithms and Applications, Springer,
2011.
2. D. Forsyth and J. Ponce, Computer Vision: A Modern Approach, Pearson
Education India, 2nd Ed, 2015.
3. R. I. Hartley and A. Zisserman, Multiple View Geometry in Computer Vision,
Cambridge University Press, 2nd Edition, 2004.
4
School of Computer Science and Engineering
Evaluation Scheme
5
School of Computer Science and Engineering
Type: Each batch will work for one distinct application area
Blooms Timeline PI
Sl. No. Experiments CO Hrs Marks
level wrt COE code
Tool Installation Assignment of
the following application areas
to each batch:
1) Semantic segmentation:
Perceptual grouping,
Agglomerative clustering
2) Motion: Optical Flow, 1st &2nd
1. CO1 L3 1.4.1 4 Nil
Pyramids for large motion. week
3) Advanced Techniques:
Image stitching, Image
pyramids, Object recognition,
Dimensionality reduction, Face
identification, Detecting objects
by parts.
2. Problem Identification CO1 L3 3rdWeek 2.1.2 2 05
Handling basic Features,
3. filtering,segmentation,clusterin L3 4th Week 2.3.1 6 Nil
CO2
g and object detection.
5th & 6th
4. Analysis CO2 L3 1.4.1 4 Nil
Week
th
5. Design CO2 L3 7 & 8 th 2.3.1 4 05
6
School of Computer Science and Engineering
Week
9th , 10th
6. Implementation CO5 L3 & 11th 5.3.1 6 10
Week
CO5 14th 10.2.2
7. Presentation and Report L3 2 10
Week 10.1.2
Total 28 30
8
School of Computer Science and Engineering
(2-3M)
Unit II
Semantic segmentation II 3 -- 0.5 -- 0.5
9
School of Computer Science and Engineering
10
School of Computer Science and Engineering
5.Apply advanced
techniques to interpret 20% ✓ ✓ ✓ ✓
results.
11
School of Computer Science and Engineering
Chapterwise Plan
Learning Outcomes:-
At the end of the topic the student should be able to:
Topic Learning Outcomes COs BL CA Code
1. Explain the concept of digital image representation. 1 L2 1.4
2. Discuss linear systems in the context of image processing. 2 L3 2.1
3. Describe convolution and cross-correlation in the context of 2 L3 2.3
image processing.
Lesson Schedule
Class No. - Portion covered per hour
1.Computer Vision Overview,
2. Pixels and image representation,
3. Filters: Linear systems,
4. Convolutions and cross-correlations;
Review Questions
Sl.No. - Questions TLOs BL PI Code
1.Explain the concept of digital image representation. How are 1 L3 1.4.1
images stored digitally? Discuss the role of pixels in representing
images and how pixel intensity contributes to image information.
2.Define linear systems in the context of image processing. Explain 2 L2 2.3.1
the properties of linearity and shift-invariance. How are linear
systems used in filtering operations in computer vision?
3.Illustrate convolution and cross-correlation in the context of 3 L3 2.1.3
image processing. Explain the mathematical formulations of both
operations and their differences
12
School of Computer Science and Engineering
Learning Outcomes:-
At the end of the topic the student should be able to:
Topic Learning Outcomes COs BL PI Code
1.Discuss the edge detection in the context of image processing. 2 L2 2.1
2.Explain the Canny edge detection algorithm. 2 L3 2.3
3. Discuss Features and fitting. 2 L2 2.3
4.Explain the Feature descriptors. 2 L3 2.3
Lesson Schedule
Class No. - Portion covered per hour
1. Edge detection: Gaussian,
2. Sobel filters,
3. Canny edge detector,
4. Features and fitting: RANSAC Local features,
5. Features and fitting: RANSAC Local features contd.
6. Harris corner detection,
7. Feature descriptors: Difference of gaussians,
8. Scale invariant feature transform;
13
School of Computer Science and Engineering
Review Questions
Sl.No. - Questions TLOs BL PI Code
1.Discuss edge detection in the context of image processing. Explain
its significance in computer vision tasks and its applications in image 1 L2 2.1.3
analysis and feature extraction.
2.Explain the Canny edge detection algorithm and its advantages
over other edge detection techniques. Discuss the steps involved in
the Canny edge detection process, including Gaussian smoothing, 2 L3 2.3.1
gradient computation, non-maximum suppression, and edge tracing
by hysteresis.
3. Discuss RANSAC (Random Sample Consensus) in the context of
computer vision. Explain its significance in robust estimation and its 3 L2 2.3.1
applications in tasks such as image stitching and object recognition.
4. Compare and contrast the performances of local feature
descriptors like SIFT and SURF with respect to robustness,
computational efficiency, and effectiveness in matching scenarios. 3 L3 2.3.1
Discuss the advantages and limitations of Harris corner detection
compared to other corner detection techniques.
5.Illustrate the Difference of Gaussians (DoG) method in the context
of computer vision. Explain how DoG is used for feature detection 4 L3 2.3.1
and its significance in scale-space representation.
14
School of Computer Science and Engineering
Learning Outcomes:-
At the end of the topic the student should be able to:
Topic Learning Outcomes COs BL PI Code
1. Explain the perceptual grouping algorithms. 3 L2 2.1
2. Discuss semantic segmentation in the context of computer vision. 3 L2 2.3
Lesson Schedule
Class No. - Portion covered per hour
1. Perceptual grouping,
2. Agglomerative clustering,
3. Super pixels and over segmentation;
Review Questions
Sl.No. - Questions TLOs BL PI Code
1.Discuss semantic segmentation in the context of computer vision.
Explain its significance in scene understanding and pixel-level 2.1.3
1 L2
classification tasks. Provide examples of real-world applications
where semantic segmentation is used.
2.Discuss the concept of perceptual grouping in image
segmentation. Explain how perceptual grouping algorithms group
image regions based on visual similarity cues such as color, texture, 1 L3 2.1.3
and proximity. Provide examples of perceptual grouping techniques
and their role in pre-processing for semantic segmentation.
15
School of Computer Science and Engineering
16
School of Computer Science and Engineering
Learning Outcomes:-
At the end of the topic the student should be able to:
Topic Learning Outcomes COs BL PI Code
1.Apply the clustering algorithms. 3 L3 2.3
2.Discuss the Visual Bag of Words (BoW) approach in the context of
3 L3 2.1
image classification
Lesson Schedule
Class No. - Portion covered per hour
1.Clustering: K-means, Mean shift;
2.Visual Bag of Words: Texture features
3.Visual bag of words…contd.
Review Questions
Sl.No. - Questions TLOs BL PI Code
1.Illustrate the clustering in the context of computer vision and its
significance in image segmentation. Explain how clustering 2.3.1
1 L3
algorithms group similar pixels together to form coherent regions in
images.
2.Describe the application of the K-means clustering algorithm in
image segmentation. Explain how K-means partitions the image
into clusters based on pixel similarities in color space. Discuss the 2.1.3
1 L3
challenges and considerations in using K-means for image
segmentation tasks, such as determining the optimal number of
clusters.
3.Explain the Visual Bag of Words (BoW) approach in the context of
image classification. Explain its significance in representing images 2 L2 2.3.1
using local features and global image descriptors.
17
School of Computer Science and Engineering
Learning Outcomes:-
At the end of the topic the student should be able to:
18
School of Computer Science and Engineering
Review Questions
Sl.No. - Questions TLOs BL PI Code
1.Discuss optical flow in the context of computer vision. Explain its
1 L2 2.1.3
significance in estimating the motion of objects in consecutive
frames of a video sequence.
2.Illustrate the Lucas-Kanade method for optical flow estimation.
Explain how it computes the motion vectors of image patches by
solving a system of linear equations using spatial and temporal 2 L3 2.1.3
gradients. Discuss its assumptions and limitations, particularly
regarding small motion and spatial smoothness
19
School of Computer Science and Engineering
Learning Outcomes:-
At the end of the topic the student should be able to:
Topic Learning Outcomes COs BL PI Code
1.Discuss image stitching in the context of computer vision 5 L2 2.1
2. Describe the concept of image pyramids and their role in 5 L3 2.4
computer vision tasks.
3. Discuss object recognition in computer vision 5 L3 2.4
Lesson Schedule
Class No. - Portion covered per hour
1.Image stitching,
2.Image pyramids ,
3.Object recognition,
4.Dimensionality reduction,
5.Face identification,
6.Detecting objects by parts
20
School of Computer Science and Engineering
Review Questions
Sl.No. - Questions TLOs BL PI Code
1.Discuss the image stitching in the context of computer vision.
Explain its significance in creating panoramic images by combining 1 L2 2.1.3
multiple overlapping images.
2.Describe the concept of image pyramids and their role in
computer vision tasks. Explain how image pyramids are constructed
by recursively downsampling images to multiple scales and how 2 L3 2.3.1
they are used for tasks such as feature detection, motion
estimation, and image blending.
3.Explain dimensionality reduction in the context of machine
learning and computer vision. Discuss techniques such as principal
component analysis (PCA) and t-distributed stochastic neighbor 3 L3 2.3.1
embedding (t-SNE) used for reducing the dimensionality of high-
dimensional feature spaces while preserving relevant information.
4.Describe face identification in computer vision. Explain its
significance in biometric authentication and surveillance systems.
Discuss techniques such as eigenfaces, local binary patterns (LBP), 4 L2 2.4.1
and deep learning-based approaches used for face identification
tasks.
21
School of Computer Science and Engineering
22
School of Computer Science and Engineering
surveillance systems.
25