Learning Path
Question & Answer1
Understand Question2
Review Options3
Learn Explanation4
Explore TopicChoose the Best Answer
A
Long Short-Term Memory (LSTM)
B
Gated Recurrent Unit (GRU)
C
Vanilla RNN
D
Feedforward Neural Network
Understanding the Answer
Let's break down why this is correct
Answer
For sentences where a word far earlier can influence a word much later, a vanilla RNN struggles because the signal fades as the sequence grows. A Long Short‑Term Memory network, or LSTM, is designed to keep a memory cell that can store information over many steps, updating it with gates that decide what to keep or forget. These gates let the model remember a word from the start of the sentence and still use it when predicting a word at the end. For example, in “The cat that chased the mouse was tired,” an LSTM can remember “cat” when deciding the adjective “tired” refers to the cat, not the mouse. Thus, LSTMs are the preferred choice for long‑term dependencies in language.
Detailed Explanation
The LSTM network uses gates that act like a smart filter, deciding what information to keep and what to drop. Other options are incorrect because The GRU is a simpler version of the LSTM and can handle some long‑term patterns, but it has fewer gates; A vanilla RNN updates its state with a single weight matrix.
Key Concepts
Recurrent Neural Networks
Long-term dependencies
Vanishing gradient problem
Topic
Recurrent Neural Networks (RNN)
Difficulty
hard level question
Cognitive Level
understand
Practice Similar Questions
Test your understanding with related questions
Ready to Master More Topics?
Join thousands of students using Seekh's interactive learning platform to excel in their studies with personalized practice and detailed explanations.