英文是纯手打的!论文原文的summarizing and paraphrasing。可能会出现难以避免的拼写错误和语法错误,若有发现欢迎评论指正!文章偏向于笔记,谨慎食用
目录
2.3. Related Work and Motivation
2.3.1. Graph Neural Networks for Graph Classification
2.3.2. Semisupervised Learning in Graph Classification
2.3.3. Few-Shot Learning in Graph Classification
2.4.1. Problem Formulation and Notations
2.4.3. Implementation on Typical Graph Classification
2.4.4. Implementation on Few-Shot Graph Classification
2.5.2. Configurations of Graph Neural Networks
2.5.3. Experiments on Typical Graph Classification
2.5.4. Experiments on Few-Shot Graph Classification
2.5.6. Parameter Sensitivity Analysis
1. 省流版
1.1. 心得
(1)还行,不算是很晦涩的论文,只是没有代码会让这个设计较难复现
2. 论文逐段精读
2.1. Abstract
①GNNs have achieve SOTA results on pure supervised area
②Semi-supervise usually applied on node classification
③They trained 2 GNNs as complementary views
2.2. Introduction
①Graph kernel methods are two stage methods and cost time
②Existing performances rely on labeled data
③They propopsed a semisupervised GNN framework for graph classification based on co-training and self-training
arduous adj.艰苦的;艰难的
2.3. Related Work and Motivation
2.3.1. Graph Neural Networks for Graph Classification
①The authors list PATCHY-SAN, MPNN, DGCNN, GAM, DIFFPOOL and GIN
②These experiments do not contain unlabeled data
2.3.2. Semisupervised Learning in Graph Classification
①Semi supervised methods: self-training, co-training, label propagation
②Assumption of semi-supervised method: smoothness assumption, the cluster assumption, and the manifold assumption
③⭐There is no connection between graphs (edge will connects nodes)
2.3.3. Few-Shot Learning in Graph Classification
①Examples a prototype model which extract the average of samples in each class:
②The probability that a graph belongs to a class:
2.4. Methodology
2.4.1. Problem Formulation and Notations
(1)Supervised Graph Classification:
①Training set:
②Mapping test data to label
(2)Semisupervised Graph Classification:
①Training set:
②Common notations:
2.4.2. Framework
①They adopt pre-trained strategy in the first epoch
②For the same graph, the 2 classifications measured by Jensen–Shannon divergence
where and denote 2 softmax score outputed by the 2 classifiers
③Supervised loss is applied on labeled data:
④Total training loss for the two GNNs:
⑤Two GNNs from different views arrange pseudo labels for each other
⑥They assign weight to each unlabeled sample:
where denotes the entropy function, is the maximum possible entropy in
⑦To weakening the impact of category imbalance, they add another weight , and the weight is defined by:
where denotes the number of labeled samples in and denotes the number of pseudo labeled samples in
⑧Loss for unlabeled data (by minimizing the pseudo labeled samples loss):
where denotes the pseudo label for unlabeled data assigned by the other view
⑨The overall loss function on co-training:
where the denotes the tradeoff factor between true labels and pseudo samples, denotes another weight for true labeled examples and
⑩They reset pseudo labeled samples every epochs to impair the harm of accumulated errors
⑪The supervised loss on pseudo labeled samples on self-training:
where denotes the pseudo label for unlabeled data assigned by their own view
⑫The overall self training loss function:
⑬The overall loss in the model:
⑭The workflow of this model:
⑮Algorithm of this model:
ameliorate vt.改善;改进;改良
2.4.3. Implementation on Typical Graph Classification
①This framework can corporate with any GNN
2.4.4. Implementation on Few-Shot Graph Classification
①They combine their framework with prototypical network:
②Framework applied in few shot classification:
③Pseudo label generation:
2.5. Experimental Study
2.5.1. Datasets
①7 classic graph classification datasets: NCI1, NC109, D&D, COLLAB, REDDIT-MULTI-12K, MiniGCDataset, and DBLP_v1. NCI1 and NCI109
②Statistics of classic graph classification datasets:
③2 small sample datasets: mini-REM12K and mini-MGCD
④Statistics of few shot graph classification datasets:
2.5.2. Configurations of Graph Neural Networks
①They chose DIFFPOOL and GIN as the two GNNs, which DIFFPOOL extracts the topological structure and GIN keeps the high order neighbor relationship
②Hyper-parameter optimization: grid search
2.5.3. Experiments on Typical Graph Classification
①Labeling rate: 0.5% and 1% on MiniGCDataset, 5% and 10% for others
②Evaluation: average performance over 10 runs
(1)Parameter Configurations
①Training epoch: 300 for original GNNs and 200 for their semisupervised GNNs
②
③
④If : , otherwise . ,
⑤
⑥
⑦Learning rate=0.001, decrese 0.5 at each 80 epochs
(2)Baseline Methods
①DIFFPOOL+ and GIN+: generated by SVM
②Strong non-GNN methods: graph2vec, Skip-Gram, RGM
(3)Results and Analysis
①Experimental results:
2.5.4. Experiments on Few-Shot Graph Classification
(1)Parameter Configurations
①
②
③ is the same as in the typical graph
④Training epoch: 40
⑤learning rate: 0.001
(2)Baseline Methods
①Similar to typical graph
(3)Results and Analysis
①Performance table:
2.5.5. Ablation Study
①Module ablation study:
2.5.6. Parameter Sensitivity Analysis
① is the number of clusters after soft coarsening in DIFFPOOL and with an increment of 0.05
② denotes the number of GNN layers and it varies from 3 to 7
③Hyperparameter varing observation:
④ on NCI1:
⑤ on mini-REM12K and mini-MGCD:
2.6. Conclusion
They want further explore the noisy labels
3. Reference
Xie, Y. et al. (2023) 'Semisupervised Graph Neural Networks for Graph Classification', IEEE Transactions on Cybernetics, 53(10): 6222-6235. doi: 10.1109/TCYB.2022.3164696