HF [Merged](https://github.com/huggingface/transformers/pull/24653) RoPE scaling into their library. This allows to increase context length by 4x without retraining.