The Transformer model, widely used for sequence modeling, relies on self-attention which becomes inefficient for long sequences.The ZETA method proposes leveraging Z-Order Curves for Efficient Top-k Attention to address the inefficiency of self-attention.ZETA enables parallel querying of past tokens and achieves similar performance to self-attention while reducing computational demands.Experimental results show that ZETA outperforms attention models on various language modeling tasks.