Cnn Archtechture
Cnn Archtechture
Answer:
AlexNet, introduced by Alex Krizhevsky et al. in 2012, revolutionized deep learning by
winning the ImageNet ILSVRC challenge with a top-5 error rate of 17% (compared
to 26% by others). Its key innovations over LeNet-5 include:
Deeper Architecture:
First CNN trained on dual NVIDIA GTX 580 GPUs (3 days vs. weeks on CPUs).
Overlapping Max-Pooling:
Activatio
Sigmoid/Tanh ReLU
n
Hardwar
CPU GPU-optimized
e
Conclusion: AlexNet’s scale and innovations (ReLU, GPU, dropout) made it the first
modern CNN, paving the way for deeper architectures.
Answer:
The Inception module is the core building block of GoogleNet (2014), which reduced
the top-5 error to 6.7% with only 6M parameters (vs. AlexNet’s 60M).
Parallel Paths:
Outputs of all paths are depth-concatenated (requires "SAME" padding for equal
width/height).
1x1 Bottlenecks:
Computational Efficiency:
Mathematical Insight:
For input tensor X of shape (H, W, C), the module computes:
text
Copy
Download
Output = Concat[
Conv1x1(X),
Conv3x3(Conv1x1(X)),
Conv5x5(Conv1x1(X)),
MaxPool3x3(X)
]
Conclusion: Inception balances depth and efficiency via bottlenecks and multi-
scale aggregation.
3. What are skip connections in ResNet? Prove mathematically why
they mitigate vanishing gradients.
(15 marks)
Answer:
Definition: Shortcut paths that add input x to the output of a layer block F(x).
Formula: H(x) = F(x) + x, where F(x) learns the residual (difference from identity).
Gradient Flow:
Mathematical Proof:
Consider a chain of L residual blocks:
Visualization:
(10 marks)
Answer:
Feature ResNet DenseNet
Feature
Single path per block All previous features reused
Reuse
Gradient
Preserved via addition Enhanced via concatenation
Flow
Key Differences:
DenseNet:
Growth rate (K): Each layer adds K new channels (e.g., K=12).
Bottlenecks: 1x1 convs compress channels before 3x3 convs
Transition Layers: Batch norm + 1x1 conv + 2x2 pooling between blocks.
ResNet:
Example:
Dense Block:
Layer1: [x]
Layer2: [x, F1(x)]
Layer3: [x, F1(x), F2(x)]
Res Block:
Layer1: x
Layer2: x + F1(x)
Conclusion: DenseNet improves feature reuse but requires more memory; ResNet is
simpler and widely adopted.