
by Rylan Schaeffer
Jason Wei presented his 2021 ICLR paper Finetuned Language Models Are Zero-Shot Learners at the Stanford NLP journal club:

The idea is simple: finetune a language model on a collection of NLP tasks, described using natural language. The result is that the model performs tasks it hasn’t seen before via instructions.

The proof


“Instruction tuning” .

62 NLP datasets, with 12 task clusters

Generate natural language instruction templates for each task

137B parameter pretrained checkpoint. Instruction tune all parameters for 30 k steps on 62 datasets, spanning 12 task clusters.

More effective on tasks where verbalizing instructions is easy

Performs better from prompt tuning


Clusters don’t have same number of datasets and different datasets have different sizes

10 templates per dataset, trying different ways of rephrasing the same goal. Additional templates has little effect

Natural language instructions critical to zero-shot learning

Contact: jasonwei@google.com
tags: natural-language-processing - deep-learning - language-models