Finetuned language models are zero-shot learners
This paper explores a simple method for improving the zero-shot learning abilities of
language models. We show that instruction tuning--finetuning language models on a�…
language models. We show that instruction tuning--finetuning language models on a�…
Multitask prompted training enables zero-shot task generalization
Large language models have recently been shown to attain reasonable zero-shot
generalization on a diverse set of tasks (Brown et al., 2020). It has been hypothesized that�…
generalization on a diverse set of tasks (Brown et al., 2020). It has been hypothesized that�…
Multitask prompted training enables zero-shot task generalization
S Victor, W Albert, R Colin, B Stephen…�- International�…, 2022 - iris.uniroma1.it
Large language models have recently been shown to attain reasonable zero-shot
generalization on a diverse set of tasks (Brown et al., 2020). It has been hypothesized that�…
generalization on a diverse set of tasks (Brown et al., 2020). It has been hypothesized that�…
Adapting language models for zero-shot learning by meta-tuning on dataset and prompt collections
Large pre-trained language models (LMs) such as GPT-3 have acquired a surprising ability
to perform zero-shot learning. For example, to classify sentiment without any training�…
to perform zero-shot learning. For example, to classify sentiment without any training�…
Pre-trained language models can be fully zero-shot learners
How can we extend a pre-trained model to many language understanding tasks, without
labeled or additional unlabeled data? Pre-trained language models (PLMs) have been�…
labeled or additional unlabeled data? Pre-trained language models (PLMs) have been�…
Prompt consistency for zero-shot task generalization
One of the most impressive results of recent NLP history is the ability of pre-trained language
models to solve new tasks in a zero-shot setting. To achieve this, NLP tasks are framed as�…
models to solve new tasks in a zero-shot setting. To achieve this, NLP tasks are framed as�…
Multiinstruct: Improving multi-modal zero-shot learning via instruction tuning
Instruction tuning, a new learning paradigm that fine-tunes pre-trained language models on
tasks specified through instructions, has shown promising zero-shot performance on various�…
tasks specified through instructions, has shown promising zero-shot performance on various�…
What language model architecture and pretraining objective works best for zero-shot generalization?
Large pretrained Transformer language models have been shown to exhibit zero-shot
generalization, ie they can perform a wide variety of tasks that they were not explicitly trained�…
generalization, ie they can perform a wide variety of tasks that they were not explicitly trained�…
Do prompt-based models really understand the meaning of their prompts?
Recently, a boom of papers has shown extraordinary progress in zero-shot and few-shot
learning with various prompt-based models. It is commonly argued that prompts help models�…
learning with various prompt-based models. It is commonly argued that prompts help models�…
Agent instructs large language models to be general zero-shot reasoners
N Crispino, K Montgomery, F Zeng, D Song…�- arXiv preprint arXiv�…, 2023 - arxiv.org
We introduce a method to improve the zero-shot reasoning abilities of large language
models on general language understanding tasks. Specifically, we build an autonomous�…
models on general language understanding tasks. Specifically, we build an autonomous�…