PRCV Viva Notes
PRCV Viva Notes
Sure! Please share the topics unit-wise, and I'll provide you with detailed notes and potential
viva questions for each.
UNIT-1
induction algorithms, rule induction, decision trees,
bayesian methods, the basic navie bayes classified, naive
bayes induction for numeric attributes, Correction to the
probability estimation. Laplace correction, No Match. Other
bayesian methods, neural networks, genetic algorithms,
instance based learning. Support vector machines
(in detail and elaborately)
1. Induction Algorithms
Definition: Algorithms used to derive general principles (rules or models) from specific
examples or data.
Types:
2. Rule Induction
Steps:
1/32
Example:
IF age < 25 AND income < 30k THEN loan approval = NO.
3. Decision Trees
Common Algorithms:
CART (Classification and Regression Trees): Works for classification and regression.
4. Bayesian Methods
P (E∣H)P (H)
P (H∣E) =
P (E)
2/32
5. Naive Bayes Classifier
Formula:
P (X∣C)P (C)
P (C∣X) =
P (X)
Convert numeric attributes into probability densities using techniques like Gaussian or
kernel density estimation.
Gaussian formula:
1 −
(x−μ)2
P (x∣C) = e 2σ 2
2πσ 2
Laplace Correction:
Formula:
nx,C + 1
P (x∣C) =
nC + ∣X∣
3/32
No Match Problem: Handles cases when no training data matches the attribute
combination. Solutions include smoothing techniques.
Bayesian Networks: Directed acyclic graphs where nodes represent variables and edges
represent dependencies.
Maximum A Posteriori (MAP): Selects the hypothesis with the highest posterior
probability.
9. Neural Networks
Components:
Layers:
Input layer.
Hidden layers.
Output layer.
Common Architectures:
Training:
4/32
10. Genetic Algorithms
Steps:
Description: Stores instances of training data and compares new inputs to them.
Algorithms:
Description: Finds the hyperplane that maximizes the margin between classes.
Kernel Trick: Maps data to higher dimensions to handle non-linearly separable data.
5/32
Viva Questions
1. Induction Algorithms:
2. Decision Trees:
3. Bayesian Methods:
4. Laplace Correction:
5. Neural Networks:
6. Genetic Algorithms:
7. Instance-Based Learning:
UNIT 2
Statistical Pattern Recognition
6/32
Features and Feature Vectors, and Classifiers
Model Complexity
Bayes' Theorem
Decision Boundaries
Parametric Methods
Steps:
1. Data collection.
7/32
2. Pre-processing.
3. Feature extraction.
4. Classification or regression.
Classification:
Regression:
Features:
Feature Vectors:
Classifiers:
8/32
4. Pre-processing and Feature Extraction
Pre-processing:
Feature Extraction:
Definition: Challenges arising when data has too many features (high dimensions).
Problems:
Solutions:
Formula:
n
y = ∑ w i xi
i=0
wi : Coefficients.
9/32
Overfitting: Too complex polynomial fits noise instead of capturing the trend.
7. Model Complexity
Key Points:
9. Bayes' Theorem
Formula:
P (E∣H)P (H)
P (H∣E) =
P (E)
Applications:
Significance:
10/32
Combines prior knowledge with evidence from data.
Examples:
Applications:
Online learning.
Algorithms:
Gradient Descent.
Expectation-Maximization (EM).
11/32
13. Linear Discriminant Functions
Equation:
g(x) = wT x + b
w: Weights.
b: Bias.
Used In: SVM, Fisher’s Linear Discriminant.
Steps:
Definition: Neural networks where data flows forward from input to output.
Structure:
Viva Questions
12/32
1. Classification and Regression:
3. Curse of Dimensionality:
5. Bayes’ Theorem:
6. Decision Boundaries:
8. Feed-forward Networks:
Let me know if you'd like more examples or deeper elaboration on any topic!
UNIT - 3
13/32
Review of image processing techniques
Thresholding techniques
Mathematical morphology
Texture
2/2
Image Processing:
Applications:
Definition: Modify image pixel values to enhance or suppress certain features (e.g.,
noise, edges).
Types:
14/32
Smoothing Filters:
2πσ 2
Sharpening Filters:
3. Thresholding Techniques
Methods:
Global Thresholding:
Local/Adaptive Thresholding:
Multilevel Thresholding:
15/32
4. Edge Detection Techniques
Methods:
Gradient-based Methods:
Formula:
2 ∂2f ∂2f
∇ f= + 2
∂x2 ∂y
Multi-stage algorithm:
3. Non-maximum suppression.
Edge Linking:
Methods:
Key equation:
R = det(M ) − k(trace(M ))2
16/32
M : Structure tensor matrix.
k : Sensitivity parameter.
Shi-Tomasi Corner Detector:
6. Mathematical Morphology
Basic Operations:
Erosion:
Dilation:
Opening:
Closing:
Applications:
17/32
Noise removal, shape analysis, skeletonization.
7. Texture
Attributes:
Texture Analysis:
Statistical Methods:
Structural Methods:
Model-based Methods:
Applications:
Viva Questions
1. Classical Filtering Operations:
2. Thresholding Techniques:
18/32
Why is the Canny edge detector considered optimal?
4. Corner Detection:
5. Mathematical Morphology:
6. Texture:
UNIT IV
Connectedness
Size filtering
Distance functions
Active contours
Centroidal profiles
19/32
Handling occlusion
Boundary descriptors
Chain codes
Fourier descriptors
Region descriptors
Moments
Goal: Extract meaningful information about the shape, structure, or form of objects.
2. Connectedness
Types:
4-connectivity: A pixel is connected to its four neighbors (up, down, left, right).
20/32
3. Object Labeling and Counting
Definition: The process of assigning unique labels to different objects in an image and
counting them.
Algorithm:
4. Size Filtering
Definition: Filters out objects based on their size (area or pixel count).
Method:
Once objects are labeled, the size of each object is calculated, and objects below or
above a certain threshold can be removed.
5. Distance Functions
Definition: Measures the distance of pixels from a given reference (e.g., boundary of an
object or another pixel).
Types:
Chamfer Distance: A fast approximation to Euclidean distance, often used for edge
tracking.
21/32
City Block Distance: Uses only vertical and horizontal movements, often in grid-
based systems.
Skeletonization:
Thinning:
Process of eroding away pixels from an object boundary until only the "skeleton"
remains.
Definition: Involves analyzing shapes that can change or deform over time, like
biological shapes or flexible structures.
Methods:
22/32
8. Boundary Tracking Procedures
Definition: The process of tracking the boundary of an object to capture its shape.
Techniques:
Edge Detection: Using algorithms like Sobel, Canny, or Prewitt to identify object
boundaries.
Contour Tracing: Traces the boundary of an object by starting from an edge pixel
and following the connected boundary pixels.
9. Active Contours
Definition: Active contours (also called snakes) are curves that move within an image to
find boundaries of objects.
Method:
Types:
Template Matching: A shape model is compared with the image to find similar
objects.
23/32
Shape Descriptors: Features (e.g., Fourier descriptors, moments) are extracted and
used for recognition.
Method:
The object’s boundary is traced from the centroid outward, capturing the object’s
characteristics.
Definition: The process of dealing with the partial blocking of an object in an image by
another object (occlusion).
Methods:
Partial Shape Matching: Matching shapes even when parts of the object are
occluded.
Methods:
24/32
Perimeter Measure: Total length of the boundary of an object.
Types:
Fourier Descriptors: Transform the boundary into a frequency domain for easier
comparison.
Properties:
25/32
Definition: A mathematical tool used to describe a shape by decomposing its boundary
into sinusoidal components.
Process:
Definition: Describes the overall shape and characteristics of an object by analyzing its
internal properties.
Types:
18. Moments
Types:
Mpq = ∑ ∑ xp y q I(x, y)
x y
26/32
Where I(x, y) is the intensity of the pixel at (x, y), and p, q are non-negative integers.
Viva Questions
1. Binary Shape Analysis:
How is object labeling performed? Can you explain the flood-fill algorithm?
3. Size Filtering:
4. Distance Functions:
6. Active Contours:
Explain the working principle of active contours (snakes) and their applications.
What are shape models and how are they used for shape recognition?
8. Chain Codes:
How do chain codes represent object boundaries? What are their advantages?
9. Fourier Descriptors:
10. Moments:
27/32
Give me more viva questions along with their answers
(from unit 1,2,3,4 (all))
2/2
Here are more Viva Questions along with answers for Units 1, 2, 3, and 4 of your Pattern
Recognition and Computer Vision syllabus:
1. What are the main steps involved in the process of inductive learning?
2. Feature Selection: Choosing relevant features that represent the data effectively.
5. Model Evaluation: Testing the model against unseen data to evaluate performance.
Answer: The Naive Bayes classifier is a probabilistic classifier based on Bayes’ Theorem,
assuming independence between features. It calculates the posterior probability of a
class based on the input features, and the class with the highest posterior probability is
chosen as the predicted class.
Answer: Laplace correction is a technique used to handle the problem of zero probability
in Naive Bayes classifiers, especially when certain feature-class combinations do not
appear in the training data. It adds a small constant (typically 1) to all observed counts to
ensure that no probability is zero.
28/32
Answer: Support Vector Machines (SVM) are supervised learning models used for
classification tasks. SVM works by finding the hyperplane that best separates the data
points of different classes, maximizing the margin between them. It uses support
vectors, the closest data points to the hyperplane, to define the margin.
Answer: Neural networks are capable of learning complex, non-linear patterns in data.
They can generalize well to unseen data, handle large and unstructured data (like
images and text), and are highly adaptable to various types of learning tasks, such as
classification, regression, and clustering.
Answer: The Curse of Dimensionality refers to the exponential growth of data volume as
the number of features or dimensions increases. In high-dimensional spaces, the
available data becomes sparse, making learning difficult and computationally expensive.
It leads to overfitting and poor model generalization.
Answer: Bayes' Theorem describes the relationship between prior probability, likelihood,
and posterior probability. In pattern recognition, it is used for classification by
calculating the posterior probability of a class given a set of features. It helps in making
predictions based on observed data and prior knowledge.
Answer: Fisher's Linear Discriminant is a method for dimensionality reduction that seeks
to find a linear combination of features that best separates two or more classes. It
maximizes the ratio of between-class variance to within-class variance, thereby achieving
better class separation.
Answer: Pre-processing and feature extraction are crucial in pattern recognition. Pre-
processing involves cleaning and transforming raw data to a suitable form for analysis.
29/32
Feature extraction reduces data dimensions by identifying the most relevant attributes,
which helps in improving the performance and efficiency of the classification models.
Answer: Edge detection techniques are used to identify boundaries of objects within an
image. Common methods include:
30/32
Answer: Corner detection is a technique to identify points in an image where two edges
meet, forming sharp angles (corners). Corners are often used for object recognition,
tracking, and matching. The Harris corner detector is a widely used algorithm for corner
detection.
Answer: Connectedness refers to how pixels are related based on their adjacency. In a
binary image, pixels that are connected by either 4 or 8 neighbors (depending on the
connectivity criterion) are considered part of the same object. Connectedness is crucial
for object labeling and segmentation.
Answer: Object labeling involves assigning a unique identifier to each distinct object or
region in an image. Counting refers to the process of determining how many objects are
present. These tasks typically involve connected component analysis, where regions of
pixels that are connected are labeled as individual objects.
Answer: Size filtering refers to the process of eliminating objects based on their size. In
practice, after detecting objects, their size (area) is computed, and objects that do not
meet a predefined size threshold (either too small or too large) are removed from
further analysis.
Answer: Skeletonization (or thinning) is the process of reducing an object to its skeletal
form, preserving its topological structure but removing unnecessary pixels. This is done
31/32
to simplify the object for further analysis, such as shape recognition or matching.
Thinning algorithms iteratively remove boundary pixels while keeping the object's
central axis intact.
Answer: Fourier descriptors are used to represent the shape of an object in the
frequency domain by decomposing its boundary into sinusoidal components. The
Fourier coefficients represent the shape, and using only a few coefficients allows for a
compact representation of the object’s boundary.
32/32