This Talk: 1) Node Embeddings 2) Graph Neural Networks 3) Applications
This Talk: 1) Node Embeddings 2) Graph Neural Networks 3) Applications
§ 1) Node embeddings
§ Map nodes to low-dimensional
embeddings.
§ 3) Applications
Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW 2018 1
Part 1:
Node Embeddings
Input Output
Intuition: Find embedding of nodes to d-
dimensions so that “similar” nodes in the graph
have embeddings that are close together.
Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW 2018 3
Setup
§ Assume we have a graph G:
§ V is the vertex set.
§ A is the adjacency matrix (assume binary).
§ No node features or extra information
is used!
Need to define!
enc(v) = Zv
d⇥|V| matrix, each column is node
Z2R embedding [what we learn!]
|V| indicator vector, all zeroes
v2I except a one in column
indicating node v
Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW 2018 9
“Shallow” Encoding
§ Simplest encoding approach: encoder
is just an embedding-lookup
embedding vector for a
embedding specific node
matrix
Dimension/size
Z= of embeddings
1. Adjacency-based similarity
2. Multi-hop similarity
3. Random walk approaches
High-level structure and material from:
• Hamilton et al. 2017. Representation Learning on Graphs:
Methods and Applications. IEEE Data Engineering Bulletin
on Graph Systems.
probability that u
> and v co-occur on
zu zv ⇡ a random walk over
the network
2. Optimize embeddings to
encode these random walk
statistics.
X X ✓ ◆
exp(z>
u zv )
L= log P >
u2V v2NR (u) n2V exp(zu zn )
X X ✓ ◆
exp(z>
u zv )
L= log P >
u2V v2NR (u) n2V exp(zu zn )
u s1
Closer to 𝒖
BFS: DFS:
Micro-view of Macro-view of
neighbourhood neighbourhood
p=1, q=2
appearance network generated by node2vec with label colors
reflecting homophily (top) and structural equivalence (bottom). p=1, q=0.5
on the following datasets:
• BlogCatalog [44]: This is a network of social relation
of the bloggers listed on the BlogCatalog website. Th
Microscopic view of the
also exclude a recent approach, GraRep [6], that generalizes LINE
Macroscopic view of the
bels represent blogger interests inferred through the
data provided by the bloggers. The network has 10,312
network neighbourhood
to incorporate information from network neighborhoods beyond 2-
hops, but does not scale and hence, provides an unfair comparison network neighbourhood
333,983 edges and 39 different labels.
• Protein-Protein Interactions (PPI) [5]: We use a sub
with other neural embedding based feature learning methods. Apart
of the PPI network for Homo Sapiens. The subgraph
from spectral clustering which has a slightly higher time complex-
responds to theWWW
graph induced by nodes for which we
ity since it involves matrix factorization, our
Representation experiments
Learning stand
on Networks, out
snap.stanford.edu/proj/embeddings-www, 2018 45
Other random walk ideas
(not covered in detail here)
§ Different kinds of biased random walks:
§ Based on node attributes (Dong et al., 2017).
§ Based on a learned weights (Abu-El-Haija et al., 2017)
§ Alternative optimization schemes:
§ Directly optimize based on 1-hop and 2-hop random
walk probabilities (as in LINE from Tang et al. 2015).
§ Network preprocessing techniques:
§ Run random walks on modified versions of the original
network (e.g., Ribeiro et al. 2017’s struct2vec, Chen et
al. 2016’s HARP).
Brain
Brainstem
Cerebellum Frontal Parietal Occipital Temporal
lobe lobe lobe lobe