Mask tokens in a sequence, and ask the model to fill them in. This differs from causal language modeling in that MLM has access to both the left and right contexts, not just the left context.
Mask tokens in a sequence, and ask the model to predict them. This differs from masked language modeling in that CLM has access only to the left context, not the right context.
Given a context, generate a continuing portion of text that coherently follows the given context. TODO: how is this different from causal language modeling?
Classify tokens according to a class e.g. a person, place or thing.
Given a text, generate a shorter text that contains the essential information.
Classify a sentence. For instance, sentiment analysis might ask whether a sentence has a positive or negative connotation, or whether two sentences are rephrasings of one another.
Also known as reading comprehension, this task entails receiving a question, reading a passage and extracting an answer to that question from the passage. Examples:
Given a sequence in one language, convert it to a sequence in another language with the same meaning.