Papers
arxiv:2206.08082

Self-Generated In-Context Learning: Leveraging Auto-regressive Language Models as a Demonstration Generator

Published on Jun 16, 2022
Authors:
,
,
,
,

Abstract

Large-scale <PRE_TAG>pre-trained language models (PLMs)</POST_TAG> are well-known for being capable of solving a task simply by conditioning a few input-label pairs dubbed <PRE_TAG>demonstrations</POST_TAG> on a prompt without being explicitly tuned for the desired downstream task. Such a process (i.e., in-context learning), however, naturally leads to high reliance on the <PRE_TAG>demonstrations</POST_TAG> which are usually selected from <PRE_TAG>external datasets</POST_TAG>. In this paper, we propose self-generated in-context learning (SG-ICL), which generates <PRE_TAG>demonstrations</POST_TAG> for in-context learning from PLM itself to minimize the reliance on the external demonstration. We conduct experiments on four different text classification tasks and show SG-ICL significantly outperforms zero-shot learning and is generally worth approximately 0.6 gold training samples. Moreover, our generated <PRE_TAG>demonstrations</POST_TAG> show more consistent performance with low variance compared to randomly selected <PRE_TAG>demonstrations</POST_TAG> from the training dataset.

Community

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2206.08082 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2206.08082 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2206.08082 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.