A framework LGKDE is proposed for learning kernel density estimation for graphs.
LGKDE leverages graph neural networks and maximum mean discrepancy to learn the graph metric for multi-scale KDE with learned parameters.
The method shows consistency and convergence guarantees, including bounds on error, robustness, and complexity.
Empirical evaluation demonstrates superior performance of LGKDE in recovering synthetic graph distributions and graph anomaly detection on benchmark datasets.