Hierarchical Graph Pooling (HGP) methods aim to address the limitations of conventional Graph Neural Networks (GNN) in terms of being inherently flat and lacking multiscale analysis.
However, most HGP methods fail to consider the global topology of the graph, focus primarily on feature learning, and align local and global features in a multiscale manner.
The proposed LGRPool method utilizes the expectation maximization framework in machine learning to align local and global aspects of message passing.
LGRPool introduces a regularizer to ensure that the global topological information is in line with the local message passing at different scales in HGP.