diff --git a/examples/alpaca/train_alpaca_lora.py b/examples/alpaca/train_alpaca_lora.py index 6872d4e..3f0d175 100644 --- a/examples/alpaca/train_alpaca_lora.py +++ b/examples/alpaca/train_alpaca_lora.py @@ -319,7 +319,6 @@ def train(load_in_8bit=False) -> None: ) # Resize the tokenizer's vocabulary size to accommodate additional special tokens, if necessary - tokenizer.pad_token = tokenizer.eos_token special_tokens_dict = {} if tokenizer.pad_token is None: special_tokens_dict['pad_token'] = DEFAULT_PAD_TOKEN