Graph foundation models using graph neural networks are being used for atomistic modeling to handle multi-source, multi-fidelity data during pre-training.
Recent studies employ multi-task learning where shared layers process atomistic structures initially regardless of the source, routing them to different decoding heads for data-specific predictions.
A new multi-task parallelism method is proposed to distribute each head across computing resources with GPU acceleration, implemented in the open-source HydraGNN architecture.
The method was trained on over 24 million structures from five datasets and tested on supercomputers like Perlmutter, Aurora, and Frontier, showing efficient scaling on heterogeneous computing architectures.