Operator learning enables data-driven modeling of partial differential equations by learning mappings between function spaces.
Mondrian introduces transformer operators that decompose a domain into non-overlapping subdomains and apply attention over sequences of subdomain-restricted functions.
This approach decouples attention from discretization and supports local and global interactions through hierarchical windowed and neighborhood attention.
Mondrian achieves strong performance on Allen-Cahn and Navier-Stokes PDEs, showcasing resolution scaling without retraining.