Transformers lack a mechanism for encoding order, but Rotary Position Embedding (RoPE) has been a popular solution for facilitating relative spatial understanding.
Scaling RoPE to handle multidimensional spatial data has been a challenge, as current designs treat each axis independently and fail to capture interdependence.
University of Manchester researchers introduced a method that extends RoPE into N dimensions using Lie group and Lie algebra theory, ensuring relativity and reversibility of positional encodings.
The method offers a mathematically complete solution, allowing for learning inter-dimensional relationships and scaling to complex N-dimensional data, improving Transformer architectures.