A collection of research papers and software related to explainability in graph machine learning. Deep learning methods are achieving ever-increasing performance on many artificial intelligence tasks. A major limitation of deep models is that they are not amenable to interpretability. This limitation can be circumvented by developing post hoc techniques to explain the predictions, giving rise to the area of explainability. Recently, explainability of deep models on images and texts has achieved significant progress. In the area of graph data, graph neural networks (GNNs) and their explainability are experiencing rapid developments. However, there is neither a unified treatment of GNN explainability methods, nor a standard benchmark and testbed for evaluations. In this survey, we provide a unified and taxonomic view of current GNN explainability methods.
Features
- Perturbation Based Methods
- Surrogate Model Based Methods
- Concept Based Methods
- Decomposition and Gradient Integration Based Methods
- Explainable Predictions
- Theory and Survey Papers