📚 Learning Guide
Degradation Problem in Deep Networks
hard

Why does increasing the depth of a neural network often lead to performance degradation despite not being caused by overfitting?

Master this concept with our detailed explanation and step-by-step learning approach

Learning Path
Learning Path

Question & Answer
1
Understand Question
2
Review Options
3
Learn Explanation
4
Explore Topic

Choose the Best Answer

A

Activation functions become non-linear, making optimization harder

B

The network learns redundant features that do not contribute to accuracy

C

The gradient can vanish or explode during backpropagation

D

The increased parameters lead to a higher training loss

Understanding the Answer

Let's break down why this is correct

Answer

Adding more layers can make it harder for the network to learn useful features because the gradient signal that trains the layers can become very weak or unstable as it travels back through many weights, a problem known as vanishing or exploding gradients; this makes the deeper layers harder to train even if the model is not overfitting. When the gradient is too small, the weights in the early layers change very little, so the network cannot improve its mapping and ends up performing worse than a shallower counterpart. Moreover, deeper architectures introduce more parameters that may not contribute to learning because the optimization landscape becomes more rugged, causing the optimizer to get stuck in suboptimal local minima. For example, a 10‑layer convolutional network on MNIST may achieve 99 % accuracy, while a 20‑layer version without special tricks can drop to 95 % simply because the deeper layers cannot be effectively trained. Residual connections or careful initialization are often needed to avoid this degradation.

Detailed Explanation

When a network is very deep, the small changes in the output are multiplied many times during backpropagation. Other options are incorrect because Activation functions are designed to add non‑linearity, but they do not make the optimization impossible; Deeper networks can learn richer features, not just useless ones.

Key Concepts

Degradation Problem in Deep Networks
Backpropagation
Neural Network Optimization
Topic

Degradation Problem in Deep Networks

Difficulty

hard level question

Cognitive Level

understand

Practice Similar Questions

Test your understanding with related questions

Ready to Master More Topics?

Join thousands of students using Seekh's interactive learning platform to excel in their studies with personalized practice and detailed explanations.