What does BERT stand for in NLP?
a) Bidirectional Encoder Representations from Transformers
b) Basic Entity Representation Technique
c) Bidirectional Embedding Representation Tool
d) Best Encoding Representation Transformer
Answer:
a) Bidirectional Encoder Representations from Transformers
Explanation:
BERT stands for Bidirectional Encoder Representations from Transformers and is a popular model used for NLP tasks like text classification and translation.
Reference:
Natural Language Processing (NLP) Quiz – MCQ Questions and Answers