Learning Path
Question & Answer1
Understand Question2
Review Options3
Learn Explanation4
Explore TopicChoose the Best Answer
A
By making the model too complex to understand
B
By causing the model to focus excessively on small variation in training data
C
By increasing the amount of data needed for training
D
By reducing the training time significantly
Understanding the Answer
Let's break down why this is correct
Answer
When gradients vanish, the model stops learning important patterns and only memorizes the training data, which makes it fit noise and overfit. When gradients explode, the weights jump wildly and the model oscillates, forcing it to chase every tiny fluctuation in the data instead of general trends. Both situations cause the model to capture idiosyncratic details that do not hold in new data, so predictions degrade on unseen business metrics. For example, a sales forecast network that keeps exploding gradients will keep adjusting to every daily price spike, learning the spike pattern rather than the underlying demand curve, and will fail when the spike pattern changes. Thus, gradient instability can turn a model into a memorizer rather than a generalizer, leading to overfitting.
Detailed Explanation
When gradients vanish, the model learns very slowly and only remembers tiny changes in the training data. Other options are incorrect because Gradient size does not make the model more complex; Vanishing or exploding gradients do not increase the amount of data needed.
Key Concepts
overfitting
Topic
Vanishing/Exploding Gradients Problem
Difficulty
easy level question
Cognitive Level
understand
Practice Similar Questions
Test your understanding with related questions
Ready to Master More Topics?
Join thousands of students using Seekh's interactive learning platform to excel in their studies with personalized practice and detailed explanations.