What is the role of attention mechanisms in NLP models?
a) To ignore less important words
b) To focus on relevant parts of the input sequence
c) To tokenize the input data
d) To summarize text
Answer:
b) To focus on relevant parts of the input sequence
Explanation:
Attention mechanisms allow NLP models to focus on important parts of the input sequence, improving tasks like translation and summarization.
Reference:
Natural Language Processing (NLP) Quiz – MCQ Questions and Answers