Finetuned language models are zero-shot learners

J Wei, M Bosma, VY Zhao, K Guu, AW Yu…�- arXiv preprint arXiv�…, 2021 - arxiv.org
This paper explores a simple method for improving the zero-shot learning abilities of
language models. We show that instruction tuning--finetuning language models on a�…

Multitask prompted training enables zero-shot task generalization

V Sanh, A Webson, C Raffel, SH Bach…�- arXiv preprint arXiv�…, 2021 - arxiv.org
Large language models have recently been shown to attain reasonable zero-shot
generalization on a diverse set of tasks (Brown et al., 2020). It has been hypothesized that�…

Multitask prompted training enables zero-shot task generalization

S Victor, W Albert, R Colin, B Stephen…�- International�…, 2022 - iris.uniroma1.it
Large language models have recently been shown to attain reasonable zero-shot
generalization on a diverse set of tasks (Brown et al., 2020). It has been hypothesized that�…

Adapting language models for zero-shot learning by meta-tuning on dataset and prompt collections

R Zhong, K Lee, Z Zhang, D Klein�- arXiv preprint arXiv:2104.04670, 2021 - arxiv.org
Large pre-trained language models (LMs) such as GPT-3 have acquired a surprising ability
to perform zero-shot learning. For example, to classify sentiment without any training�…

Pre-trained language models can be fully zero-shot learners

X Zhao, S Ouyang, Z Yu, M Wu, L Li�- arXiv preprint arXiv:2212.06950, 2022 - arxiv.org
How can we extend a pre-trained model to many language understanding tasks, without
labeled or additional unlabeled data? Pre-trained language models (PLMs) have been�…

Prompt consistency for zero-shot task generalization

C Zhou, J He, X Ma, T Berg-Kirkpatrick…�- arXiv preprint arXiv�…, 2022 - arxiv.org
One of the most impressive results of recent NLP history is the ability of pre-trained language
models to solve new tasks in a zero-shot setting. To achieve this, NLP tasks are framed as�…

Multiinstruct: Improving multi-modal zero-shot learning via instruction tuning

Z Xu, Y Shen, L Huang�- arXiv preprint arXiv:2212.10773, 2022 - arxiv.org
Instruction tuning, a new learning paradigm that fine-tunes pre-trained language models on
tasks specified through instructions, has shown promising zero-shot performance on various�…

What language model architecture and pretraining objective works best for zero-shot generalization?

T Wang, A Roberts, D Hesslow…�- International�…, 2022 - proceedings.mlr.press
Large pretrained Transformer language models have been shown to exhibit zero-shot
generalization, ie they can perform a wide variety of tasks that they were not explicitly trained�…

Do prompt-based models really understand the meaning of their prompts?

A Webson, E Pavlick�- arXiv preprint arXiv:2109.01247, 2021 - arxiv.org
Recently, a boom of papers has shown extraordinary progress in zero-shot and few-shot
learning with various prompt-based models. It is commonly argued that prompts help models�…

Agent instructs large language models to be general zero-shot reasoners

N Crispino, K Montgomery, F Zeng, D Song…�- arXiv preprint arXiv�…, 2023 - arxiv.org
We introduce a method to improve the zero-shot reasoning abilities of large language
models on general language understanding tasks. Specifically, we build an autonomous�…