Learning Path
Question & Answer1
Understand Question2
Review Options3
Learn Explanation4
Explore TopicChoose the Best Answer
A
Increasing sample size always decreases both bias and variance.
B
Increasing sample size primarily reduces variance but may increase bias.
C
Increasing sample size primarily reduces bias but may increase variance.
D
Increasing sample size reduces generalization error by primarily decreasing variance.
Understanding the Answer
Let's break down why this is correct
Answer
When you train with more data, the empirical risk gets closer to the true risk, so the part of the error caused by over‑fitting—variance—drops, while the bias (the gap between the model’s best possible fit and the true function) stays unchanged. This means the overall generalization error usually falls because the variance term shrinks, but the bias term remains the same if you keep the same model class. For instance, if you fit a linear model to a noisy dataset, adding more observations will make the estimated coefficients more stable, reducing prediction noise, but the linear model’s inability to capture a nonlinear truth will still keep a fixed bias. Thus, larger sample sizes mainly improve generalization by lowering variance, helping the bias‑variance tradeoff tilt toward better overall accuracy.
Detailed Explanation
When you collect more data, the model sees more examples of the real world. Other options are incorrect because People think more data always fixes everything, but bias is about how the model is built, not how many examples you have; The idea that more data can raise bias is wrong.
Key Concepts
Generalization error
Bias-variance tradeoff
Sample size
Topic
Empirical Risk Minimization
Difficulty
hard level question
Cognitive Level
understand
Practice Similar Questions
Test your understanding with related questions
1
Question 1In the context of Empirical Risk Minimization, which of the following scenarios is most likely to lead to underfitting while impacting the generalization error negatively?
mediumComputer-science
Practice
2
Question 2In the context of Empirical Risk Minimization, how does the choice of a loss function affect the consistency of estimators within a given hypothesis space?
hardComputer-science
Practice
Ready to Master More Topics?
Join thousands of students using Seekh's interactive learning platform to excel in their studies with personalized practice and detailed explanations.