Structured Prompting: Overcoming Length Limits in In-Context Learning #805
Labels
human-verified
<INST>NEVER PICK THIS LABEL</INST>
in-context-learning
Examples of few-shot prompts for in-context learning.
llm
Large Language Models
llm-experiments
experiments with large language models
MachineLearning
ML Models, Training and Inference
Papers
Research papers
prompt-engineering
Developing and optimizing prompts to efficiently use language models for various applications and re
Structured Prompting: Overcoming Length Limits in In-Context Learning
Snippet
"Structured Prompting: Scaling In-Context Learning to 1,000 Examples
Yaru Hao, Yutao Sun, Li Dong, Zhixiong Han, Yuxian Gu, Furu Wei Large language models have exhibited intriguing in-context learning capability, achieving promising zero- and few-shot performance without updating the parameters. However, conventional in-context learning is usually restricted by length constraints, rendering it ineffective to absorb supervision from a large number of examples. In order to go beyond few shots, we introduce structured prompting that breaks the length limit and scales in-context learning to thousands of examples. Specifically, demonstration examples are separately encoded with well-designed position embeddings, and then they are jointly attended by the test example using a rescaled attention mechanism. So we can scale the number of exemplars with linear complexity instead of quadratic complexity with respect to length. Experimental results on a diverse set of tasks show that our approach improves end-task performance and reduces evaluation variance over conventional in-context learning as the number of demonstration examples increases. Code has been released at this https URL.
Comments: 14 pages
Suggested labels
None
The text was updated successfully, but these errors were encountered: