论文网址:[2010.13663] Contrastive Graph Neural Network Explanation (arxiv.org)
论文代码:GitHub - lukasjf/contrastive-gnn-explanation
英文是纯手打的!论文原文的summarizing and paraphrasing。可能会出现难以避免的拼写错误和语法错误,若有发现欢迎评论指正!文章偏向于笔记,谨慎食用
目录
1. 心得
(1)正文只有四页!冲刺!
(2)一个非常简单的损失~
2. 论文逐段精读
2.1. Abstract
①They think occlusion fail for one elimination results large difference
②They called the situation as Distribution Compliant Explanation (DCE), and they only use data consistent with the training distribution for model interpretation
③They proposed a Contrastive GNN Explanation (CoGE) technique
2.2. Introduction
①⭐Occlusion can be used in GNN explaination, but it's too extreme that one node might greatly change the structure of a sparse graph
②⭐Excluding one edge in a graph may cause disconnective graph
③⭐CoGE searches the similarity between nodes in the same label and dissimilarity between nodes in different label
④Edge explaination methods:
2.3. Related work
(1)Graph Neural Networks
(2)Explainability Methods for Graphs
(3)Adversarial Graph Attacks
2.4. Method
(1)Preliminaries
①Considering undirected graph with node set and edge set
②Feature matrix
(2)Explanations for graph classification
①They measure the similarity by Optimal Transport (OT) distance
②A example of how to calculate the distance between left graph and the middle graph:
each node holds a weight and all the weigts in one graph equals to 1. The capacity of one node is the weight. The cost of transport is the source weight multiples the distance (L2 distance)
③They aim to find a weight:
(第一项的上标其实是“不约等于”,但csdn怎么没有这个标志?)
where the first term means the average distance of the most similar graph with different label, the second term is the average distance of the most similar graph with the same label, the third term is the distance between weighted graph and its uniformly-weighted version.
④The formal loss:
(将第一个公式中的两个“不等于”替换成“不约等于”)
2.5. Experiments
2.5.1. CoGE Implementation
①Number of compared graphs:
②Optimizer: Adam
③Learning rate: 0.1, only 0.01 for REDDIT
2.5.2. Qualitative Analysis
①Graph classification dataset: MUTAG (4337 chemical molecules) and REDDIT-BINARY (2000 Reddit threads)
②GNN: GIN
③The most important structure in MUTAG:
where the left denotes the original graph, the middle denotes the similar graph with the same label, the right one is the similar graph with different label
④The most important structure in REDDIT-BINARY:
where the number denotes the degree
2.5.3. Quantitative Analysis
(1)Dataset
①Node classification dataset: CYCLIQ
②Aiming: finding how many of the most important edges are in the loop or cluster
(2)Experiment Setup
①GNN: GCN with 5 layers
②Embedding size: 20
③Edge features: NONE
④Split: 80%/20% train/test
(3)Results
①Performance on CYCLIQ dataset:
(4)AblationStudy
①Loss ablation:
and they also tried euclidean distance on the weighted average on the node embeddings (L and Average) and got a worse result
2.6. Conclusion
They aim to further apply it in node classification
3. Reference
Faber, L., Moghaddam, A. K., & Wattenhofer, R. (2020) 'Contrastive Graph Neural Network Explanation', ICML Workshop on Graph Representation Learning and Beyond. doi: https://doi.org/10.48550/arXiv.2010.13663