Ishigurognnintroduction201023 201027054344
Ishigurognnintroduction201023 201027054344
Ishigurognnintroduction201023 201027054344
1
Take home message
• Graph Neural Networks (GNNs): Neural Networks (NNs) to
compute nodes’ representation of graph-structured data
2
Table of Contents (1/2)
3
Table of Contents (2/2)
5
Graph: for relational data
• Graph
• Set of nodes (vertices): :
• Set of edges:
– directional/non-directional
Adjacency Matrix 0 1 1 0 0
1 0 0 1 1
1 0 0 0 1
Edge existence between
N X N node pairs 0 1 0 0 0
0 1 1 0 0
Symmetric
7 A <-- non-directional edges
GNN Functionality: compute latent representation
H
Latent repr. Matrix H of nodes
Node
Prediction
tasks
L-layered
GNN
0 1 1 0 0
1 0 0 1 1
1 0 0 0 1 Latent repr. Vector h_G of a graph
0 1 0 0 0 Graph
0 1 1 0 0 Prediction
tasks
8
Interface for intelligent processes on graphs
L-layered Regression
GNN
Interpretation
Informative features
ML-friendly structure (set of vectors)
Generative
9 models
Application Domains of GNNs: concrete
applications
Knowledge Graph
Social Network Chemical Molecular Graph [DeCao19QA2]
[Jin18JTVAE, Laugier18CGCNN]
[http://konect.uni-
koblenz.de/networks/ucidata-
zachary]
10
GNN in Machine Learning Researches
~2016: “graphical models”, “graph kernels”, and pure graph theory works
2017~: GNN works add up the counts
f: Sound signal
g: Impulse response
Hall resonated sound
(reverb) abcpedia.acoustics.jp
13
E.g. ConvNet in computer vision [Krizhevsky12ConvNet]
2012 ILSVRC competition
CnvNet (Alexnet)
[Krizhevsky12ConvNet]
10% margin
14 [http://image-net.org/challenges/LSVRC/2012/ilsvrc2012.pdf]
E.g. ConvNet in computer vision [Krizhevsky12ConvNet]
Layer l
Layer l-1
shift signal filter
[Kipf18talk]
15
How can we define Conv. in graphs?
16 [いらすとや]
Fourier Transform and Convolution
Convolutions in the signal (spatial) domain = Multiplication
in the spectral (frequency) domain
1D case:
[Ma20tutorial]
18
Def: Graph Laplacian and Normalized Laplacian
2 -1 -1 0 0
Graph Laplacian Matrix -1 3 0 -1 -1
-1 0 2 0 -1
0 -1 0 1 0
0 1 1 0 0
Adjacency Matrix 0 -1 -1 0 2
1 0 0 1 1
1 0 0 0 1
Normalized Graph Laplacian
0 1 0 0 0
0 1 1 0 0 1 0 0
2 0 0 0 0 Degree Matrix 1 0
0 3 0 0 0
0 1 0
0 0 2 0 0
0 0 0 1 0 0 -1 0 1 0
0 0 0 0 2
19 0 0 1
Graph Fourier Transform =
Multiply Eigenvectors of (Normalized) Graph
Laplacians
i-th eigenvevtor i-th eigenvalue
= i-th graph Fourier base = frequency of i-th base
20
Graph Fourier Transform: just multiply U^T or U
GFT
[Ma20tutorial,Shuman13Graph]
21
Graph Convolution via GFT
Convolution = simple
multiplication in spectral
22
Naïve Graph Convolution
Two parameters: Fixed (eigenvectors of the Laplacian = Fourier bases over freq.)
Trainable
Chebychev
Polynomial
24
No U-s = No Eig. Decomp.
26
spatial interpretation of GCN’s convolution
27
GCN formulation: multi-dim simpler conv +
activation
Non-linear activation (e.g. sigmoid, ReLU)
Update rule of GCN -th layer
28
Once trained, run anywhere (on different
topologies)
• The sole trainable parameter W does not care the
adjacency
Applicable for
Trained GCN
Different Graphs
GCN
いらすとや
https://www.irasutoya.com 29
The simplest GCN… it works quite well
30
Exemplar task: semi-supervised node
classification with GCN
Number of Nodes: N (many) Set of labeled nodes (supervisions)
33
Prediction using the Trained Model
``unlabeled’’ Predict yj (a label of node j) in unlabeled set,
perform forward path from the observed xj
34
Application in Chemo-
Informatics: Protein
Interface Prediction
35
Interacting proteins
Binding sites:
Ligand interface for connection
Drug A
human organ, cell, ….
Drug A’
Receptor
36
Predict binding site nodes by GNN [Fout17Interface]
[Fout17Interface]
40
GCN formulation choices [Fout17Interface]
Considering nodes
features only
With ordering of
nearest node orders
[Fout17Interface]
42
[Fout17Interface]
43
Computer Vision
Application Example:
Scene Graph generation
Scene Graph: summary graph of image contents
[Qi19AttentiveRN]
45
Graph R-CNN for scene-graph generation
[Yang18GRCNN]
• Consists of three inner components
– (a) (b) Off-the-shelf object detector [Ren15RCNN] (omit)
– (b) (c) RePN to prune unnecessary branches (omit)
– (c) (d) Attentional GCN to predict labels of nodes and edges
46 [Yang18GRCNN]
One GNN can infer
Expand Graph with Relation Nodes all objects and relations
Graph
Expansion
Objects detected
[Yang18GRCNN]
Attentional GCN
Attention conncection
Relation node’s
latent vector
Parameter f can switch based on node types
48
Inferring labels of objects/relations
Expanded Graph Ground Truth Scene Graph
49 [Yang18GRCNN]
predicted predicted
predicted
predicted
predicted predicted
50
[Yang18GRCNN]
Other Application of GNNs
51
Theoretical issues
52
Theoretical Topics of GNN
53
``Deep” GNNs do not work well
Quite different from the successful deep Conv models in Computer vision
Better
[Kipf_Welling17GCN]
54
Oversmoothing Problem [Li18Oversmoothing]
• Latent node vectors get closer to each other as the
GCN layers go deeper (higher)
• Difficult to distinguish nodes in deeper GCNs
Latent node vectors of ``Karate club’’ social network [Li18Oversmoothing]
55
Oversmoothing is theoretically proven
• Deeper GCN converges to a solution where
connected nodes will have similar latent vectors
[Li18Oversmoothing, NT_Maehara19revisit]
Residual term:
Add the current layer ‘’as it is’’
58
Graph isomorphism problem (for theory of GNNs)
The same
graph?
[ https://ja.wikipedia.org/wiki/グラフ同型 ]
59
Weisfeiler-Lehman (WL) test algorithm [Weisfeiler_Lehman68WL]
60
Weisfeiler-Lehman (WL) test algorithm [Weisfeiler_Lehman68WL]
[Shervashidze11WLKernel]
61
Upper limit of the GNN’s representation power
62
Graph Isomorphism Network (GIN) [Xu19GIN]
A layer update of
typical GNNs
A layer update of
the proposed GIN
[Maron19NeurIPS]
64
More powerful GNNs [Sato20Survey]
65
Conclusion and
Materials
66
Take home message (Revisited)
• Graph Neural Networks (GNNs): Neural Networks (NNs) to
compute nodes’ representation of graph-structured data
67
68
Surveys and documents for studying GNN
• MoleculeNet [Wu18MoleculeNet]
– Molecular graph datasets from several chemical field.
TensorFlow implmenetaions attached
• Scene Graphs
– Visual Genome[Krishna17VG]: 108K images
• Pointcloud
– S3DIS [Armeni16S3DIS]: inside office
72
Programing Libraries
76
References M-P
[Ma20tutorial] Ma+, “Graph Neural Networks: Models and Applications”, AAAI, 2020.
http://cse.msu.edu/~mayao4/tutorials/aaai2020/
[Madhawa19GNVP] Madhawa+, “GraphNVP: An Invertible Flow Model for Generating Molecular Graphs”, arXiv:
1905.11600, 2019.
[Marrn19NeurIPS] Marron+, “Provably Powerful Graph Networks”, NeurIPS, 2019.
[Maron19ICLR] Marron+, “Invariant and Equivariant Graph Networks”, ICLR 2019.
[Morris19kGNN] Morris+, “Weisfeiler and Lehman Go Neural: Higher-order Graph Neural Networks”, AAAI, 2019.
[Ngueyn_Maehara20Homomorph] Nguyen and Maehara, “Graph Homomorphism Network”, ICML, 2020.
[NT_Maehara19revisit] NT and Maehara, “Revisiting Graph Neural Networks: All We Have is Low-pass Filters”, arXiv:
1905.09550, 2019.
[Oono_Suzuki20Exponential] Oono and Suzuki, “Graph Neural Networks Exponentially Lose Expressive Power for
Node Classification”, ICLR 2020.
[Pei20GeomGCN] Pei+, “Geom-GCN: Geometric Graph Convolutional Networks”, ICLR, 2020.
77
References Q-V
[Qi19AttentiveRN] Qi+, “Attentive Relataional Networks for Mapping Images to Scene Graphs”, CVPR, 2019.
[Sanchez-Gonzalez20GNS] Sanchez-Gonzalez+, “Learning to Simulate Complex Physics with Graph Networks”,
arXiv: 2002.09405. 2020.
[Sanyal20PGCN] Sanyal+, “Protein GCN: Protein model quality assessment using graph convolutional networks”,
bioRxiv, 2020.
[Sato19CPNGNN] Sato+, “Approximation Ratios of Graph Neural Networks for Combinatorial Problems”, NeurIPS,
2019.
[Sato20Survey] Sato, “A Survey on the Expressive Power of Graph Neural Networks”, arXiv: 2003.04078.
[Scarselli08GNN] Scarselli+, “The Graph Neural Network Model”, IEEE Trans. Neural Networks, 20(1), pp. 61-80,
2008
[Schlichtkrull18RGCN] Modeling Relational Data with Graph Convolutional Networks, ESWC 2018.
[Shervashidze11WLKernel: Shervashidez+, “Weisfeiler-Lehman Graph Kernels”, JMLR, 12, pp.2539-2661, 2011.
[Shuman13Graph] Shuman+, “The emerging field of signal processing on graphs: extending high-dimensional data
analysis to networks and other irregular domains”, IEEE Signal Processing Magazine, 30(3), pp.83-98, 2013
[Togninalli18WWL] Togninalli+, “Wasserstein Weisfeiler-Lehman Graph Kernels”, NeurIPS 2018.
[Veličković18GAT] Veličković+, “Graph Attention Networks”, ICLR 2018.
References W-Z
[Wang18NLNN] Wang+, “Non-local Neural Networks”, CVPR 2018.
[Weisfeiler_Lehman68WL] Weisfeiler and Lehman, “A Reduction of a Graph to a Canonical Form and an Algebra
Arising during this Reduction”, Nauchno-Technicheskaya Informatsia, Ser.2(9), pp.12-16, 1968.
[Wu18MoleculeNet] Wu+, “MoleculeNet: a benchmark for molecular machine learning”, Chemical Science, 9(513),
2018.
[Wu19SGC] Wu+, “Simplifying Graph Convolutional Networks”, in Proc. ICML, 2019.
[Wu19survey] Wu+, “A Comprehensive Survey on Graph Neural Networks”, arXiv:1901.00596v1 [cs.LG], 2019.
[Xu19GIN] Xu+, “How powerful are Graph Neural Networks?”, in Proc. ICLR, 2019.
[Yang18GRCNN] Yang+, “Graph R-CNN for Scene Graph Generation”, ECCV, 2018.
[You18GCPN] You+, “Graph Convolutional Policy Network for Goal-Directed Molecular Graph Generation”, NeurIPS
2018.
[Zhao_Akoglu20PairNorm] Zhao and Akoglu, “PairNorm: Tackling Oversmoothing in GNNs”, ICLR, 2020.
[Zhou20Deep] Zhou+, “Effective Training Strategies for Deep Graph Neural Networks”, arXiv: 2006.07107, 2020.
[Zhou18survey] Zhou+, “Graph Neural Networks: A Review of Methods and Applications”, arXiv: 1812.08434v2
[cs.LG], 2018. 79
GNN research history
Model Research Deeper theory
Birth of GNN The Door Opener GNN extremes
Explosion / applications