Fill-Mask
Masked language modeling is the task of masking some of the words in a sentence and predicting which words should replace those masks. These models are useful when we want to get a statistical understanding of the language in which the model is trained in.
Input
The <mask> barked at me
About Fill-Mask
Use Cases
Domain Adaptation 👩⚕️
Masked language models do not require labelled data! They are trained by masking a couple of words in sentences and the model is expected to guess the masked word. This makes it very practical!
For example, masked language modeling is used to train large models for domain-specific problems. If you have to work on a domain-specific task, such as retrieving information from medical research papers, you can train a masked language model using those papers. 📄
The resulting model has a statistical understanding of the language used in medical research papers, and can be further trained in a process called fine-tuning to solve different tasks, such as Text Classification or Question Answering to build a medical research papers information extraction system. 👩⚕️ Pre-training on domain-specific data tends to yield better results (see this paper for an example).
If you don't have the data to train a masked language model, you can also use an existing domain-specific masked language model from the Hub and fine-tune it with your smaller task dataset. That's the magic of Open Source and sharing your work! 🎉
Inference with Fill-Mask Pipeline
You can use the 🤗 Transformers library fill-mask
pipeline to do inference with masked language models. If a model name is not provided, the pipeline will be initialized with distilroberta-base. You can provide masked text and it will return a list of possible mask values ranked according to the score.
from transformers import pipeline
classifier = pipeline("fill-mask")
classifier("Paris is the <mask> of France.")
# [{'score': 0.7, 'sequence': 'Paris is the capital of France.'},
# {'score': 0.2, 'sequence': 'Paris is the birthplace of France.'},
# {'score': 0.1, 'sequence': 'Paris is the heart of France.'}]
Useful Resources
Would you like to learn more about the topic? Awesome! Here you can find some curated resources that can be helpful to you!
- Course Chapter on Fine-tuning a Masked Language Model
- Workshop on Pretraining Language Models and CodeParrot
- BERT 101: State Of The Art NLP Model Explained
- Nyströmformer: Approximating self-attention in linear time and memory via the Nyström method
Notebooks
- Pre-training an MLM for JAX/Flax
- Masked language modeling in TensorFlow
- Masked language modeling in PyTorch
Scripts for training
Documentation
Compatible libraries
Note A faster and smaller model than the famous BERT model.
Note A multilingual model trained on 100 languages.
Note A common dataset that is used to train models for many languages.
Note A large English dataset with text crawled from the web.
No example Space is defined for this task.
Note Contribute by proposing a Space for this task !
- cross_entropy
- Cross Entropy is a metric that calculates the difference between two probability distributions. Each probability distribution is the distribution of predicted words
- perplexity
- Perplexity is the exponential of the cross-entropy loss. It evaluates the probabilities assigned to the next word by the model. Lower perplexity indicates better performance