Skip to main content

Masked Language Modeling

Definition

Masked Language Modeling is a self-supervised learning task used to pre-train transformer models for natural language processing. During training, certain tokens in a sentence are intentionally hidden or “masked,” and the model is tasked with predicting these missing tokens based on their context. This process enables the model to learn deep contextual representations of words and sentences. It forms the foundation for many state-of-the-art language models.