Papers
arxiv:2206.12131

MVP: Multi-task Supervised Pre-training for Natural Language Generation

Published on Jun 24, 2022
Authors:
,
,
,

Abstract

A multi-task supervised pre-training approach for natural language generation improves performance across diverse tasks by using a unified text-to-text format and task-specific soft prompts.

AI-generated summary

Pre-trained language models (PLMs) have achieved remarkable success in natural language generation (NLG) tasks. Up to now, most NLG-oriented PLMs are pre-trained in an unsupervised manner using the large-scale general corpus. In the meanwhile, an increasing number of models pre-trained with labeled data (i.e., ``supervised pre-training'') showcase superior performance compared to unsupervised pre-trained models. Motivated by the success of supervised pre-training, we propose Multi-task superVised Pre-training~(MVP) for natural language generation. We collect a large-scale natural language generation corpus, MVPCorpus, from 77 datasets over 11 diverse NLG tasks. Then we unify these examples into a general text-to-text format to pre-train the text generation model MVP in a supervised manner. For each task, we further pre-train specific soft prompts to stimulate the model's capacity to perform a specific task. Extensive experiments have demonstrated the effectiveness and generality of our MVP model in a number of NLG tasks, which achieves state-of-the-art performance on 13 out of 17 datasets.

Community

Sign up or log in to comment

Get this paper in your agent:

hf papers read 2206.12131
Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash

Models citing this paper 16

Browse 16 models citing this paper

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2206.12131 in a dataset README.md to link it from this page.

Spaces citing this paper 1

Collections including this paper 2