This directory contains template styles for the prompts used to finetune LoRA models.
A template is described via a JSON file with the following keys:
prompt_input
: The template to use when input is not None. Uses{instruction}
and{input}
placeholders.prompt_no_input
: The template to use when input is None. Uses{instruction}
placeholders.description
: A short description of the template, with possible use cases.response_split
: The text to use as separator when cutting real response from the model output.
No {response}
placeholder was used, since the response is always the last element of the template and is just to be concatenated to the rest.
The default template, used unless otherwise specified, is alpaca.json
{
"description": "Template used by Alpaca.",
"prompt_input": "Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.\n\n### Instruction:\n{instruction}\n\n### Input:\n{input}\n\n### Response:\n",
"prompt_no_input": "Below is an instruction that describes a task. Write a response that appropriately completes the request.\n\n### Instruction:\n{instruction}\n\n### Response:\n",
"response_split": "### Response:"
}
Default template used for generic LoRA fine tunes so far.