This repository contains various experiments in extending the effective context length of models trained with Rotary Positional Embeddings. Formalized results will be provided in an upcoming paper. In the meantime, have fun with it 🙂