Learning Path
Question & Answer1
Understand Question2
Review Options3
Learn Explanation4
Explore TopicChoose the Best Answer
A
Knowledge Transfer
B
Sentence Similarity
C
Feature Extraction
D
Data Compression
Understanding the Answer
Let's break down why this is correct
Answer
Attention mechanisms help systems find the most relevant pieces of information, much like a search engine pulls up the best documents. Contextual embeddings do a similar job but for language: they turn each word into a vector that depends on its surroundings, so the model knows exactly what that word means in that sentence. These embeddings are then used for a wide range of language tasks, such as translating a sentence, answering a question, or deciding if a review is positive or negative. For example, the word “bank” in “river bank” gets a different embedding than in “bank account,” allowing the model to pick the right meaning.
Detailed Explanation
Contextual embeddings turn words into points in a space. Other options are incorrect because A common mistake is to think embeddings carry knowledge to new tasks; People think embeddings simply pull out features like a filter.
Key Concepts
Attention Mechanisms
Contextual Embeddings
Semantic Relationships
Topic
Attention Mechanisms
Difficulty
easy level question
Cognitive Level
understand
Ready to Master More Topics?
Join thousands of students using Seekh's interactive learning platform to excel in their studies with personalized practice and detailed explanations.