An Experimental Evaluation of Similarity-Based and Embedding-Based Link Prediction Methods On Graphs
An Experimental Evaluation of Similarity-Based and Embedding-Based Link Prediction Methods On Graphs
5, September 2021
An experimental evaluation of
similarity-based and embedding-based link
prediction methods on graphs
ABSTRACT
The task of inferring missing links or predicting future ones in a graph based on its current structure
is referred to as link prediction. Link prediction methods that are based on pairwise node similarity
are well-established approaches in the literature and show good prediction performance in many real-
world graphs though they are heuristic. On the other hand, graph embedding approaches learn low-
dimensional representation of nodes in graph and are capable of capturing inherent graph features,
and thus support the subsequent link prediction task in graph. This paper studies a selection of
methods from both categories on several benchmark (homogeneous) graphs with different properties
from various domains. Beyond the intra and inter category comparison of the performances of the
methods, our aim is also to uncover interesting connections between Graph Neural Network(GNN)-
based methods and heuristic ones as a means to alleviate the black-box well-known limitation.
KEYWORDS
Link Prediction, Graph Neural Network, Homogeneous Graph & Node Embedding
1 INTRODUCTION
One of the most interesting and long-standing problems in the field of graph mining is link
prediction that predicts the probability of a link between two unconnected nodes based on
available information in the current graph such as node attributes or graph structure [1].
The prediction of missing or potential links helps us toward the deep understanding of
structure, evolution and functions of real-world complex graphs [2]. Some applications of
link prediction include friend recommendation in social networks [3], product recommen-
dation in e-commerce [4], and knowledge graph completion [5].
A large category of link prediction methods is based on some heuristics that measure the
proximity between nodes to predict whether they are likely to have a link. Though these
heuristics can predict links with high accuracy in many graphs, they lack universal appli-
cability to any kind of graphs. For example, the common neighbor heuristic assumes that
two nodes are more likely to connect if they have many common neighbors. This assump-
tion may be correct in social networks, but is shown to fail in protein-protein interaction
DOI:10.5121/ijdkp.2021.11501 1
1
International Journal of Data Mining & Knowledge Management Process (IJDKP), Vol.11, No.5, September 2021
(PPI) networks [6]. In case of using these heuristics, it is required to manually choose
different heuristics for different graphs based on prior beliefs or rich expertise.
On the other hand, machine learning methods have shown their impressive performance
in many real-world applications like image classification, natural language processing etc.
The built models assume that the input data is represented as independent vectors in
a vector space. This assumption is no longer applicable for graph data as graph is a non-
Euclidean structure and the nodes in a graph are linked to some other nodes [7]. To
overcome this limitation, a lot of efforts have been devoted to develop novel graph embed-
dings where the nodes, edges, graphs are represented in a low-dimensional vector space.
In last decade, graph embedding has been established as a popular supporting tool for
solving several analytical problems in graphs like node classification, node clustering, link
prediction. The embedding approaches represent a part of a graph (or the whole graph)
in a low dimensional vector space while preserving the graph information [8].
There are some review studies in the literature which focus either on similarity-based ap-
proaches [9], [10] or embedding-based approaches [8], [11] for link prediction task in graphs.
Thus, to the best of our knowledge, a study including methods from both categories is
missing in the literature. In this paper, we try to fill this gap. This is an extended version
of our conference paper [12]. We first introduce the link prediction problem and briefly
describe selected similarity-based and embedding-based methods. Then, we evaluate their
performances on different types of graphs, namely homogeneous graphs. We compare their
performances on diverse graph groups (sharing characteristics). We also propose a few in-
teresting connections between similarity-based and embedding-based methods. Finally,
we list some future research works.
Several link prediction approaches have been proposed in the literature. We focus on
the two popular categories: (1) similarity-based approaches and (2) embedding-based
approaches.
2
International Journal of Data Mining & Knowledge Management Process (IJDKP), Vol.11, No.5, September 2021
tion) [17] use node degree, common neighborhood or links among common neighborhood
3
3
International Journal of Data Mining & Knowledge Management Process (IJDKP), Vol.11, No.5, September 2021
information to compute similarity scores. AA, RA and CCLP handcraft the computation
of weight of each common neighbours based on their neighbourhood size or clustering co-
efficient(CC). The clustering coefficient is defined as the ratio of the number of triangles
and the expected number of triangles passing through a node [17]. If tz is the number
of triangles passing through node z and Γz is the neighbourhood of z then the clustering
coefficient (CCz) of node z is defined in Equation 1.
2 × tz
CCz = (1)
|Γz| (|Γz| − 1)
On the other hand, HPI, PA and LLHN assigns equal weights to neighbours. These local
similarity-based approaches except PA work well when the graphs have a high number of
common neighbours between a pair of nodes. However, LLHN suffers from outlier (infinite
similarity score) when one of the end nodes has no neighbour. HPI also suffers from the
outlier (infinite similarity score) when both of end nodes have no neighbour.
Node2Vec:
Motivated by the classical skip-gram model in natural language processing, Grover&
Leskovec [19] developed Node2Vec representation learning approach that optimizes a
neighbourhood preserving objective function using Stochastic Gradient Descent (SGD).
Node2Vec starts with a fixed size neighbourhood sampling using guided random walk.
Unlike the classical random walk, Node2Vec defines a 2 nd order random walk that inter-
polate between BFS(Breadth First Search) and DFS(Depth First Search)-based sampling
4
4
International Journal of Data Mining & Knowledge Management Process (IJDKP), Vol.11, No.5, September 2021
strategy where two parameters p and q are used to compute the transition probability
during the walk. These parameters control how fast the walk explores and leaves the
neighborhood of the starting node. The node embedding is then generated based on the
popular skip-gram model where the co-occurrence probability among the neighbours those
appear within a window.
As illustrated in Fig. 2, WLNM is a three steps link prediction approach that starts with
extracting sub-graphs those contain a predefined number of neighbour nodes, labelling
and encoding the nodes in the sub-graph using WL algorithm and ends with training and
evaluating the neural network. WLNM is a simple GNN-based link prediction approach
which is able to learn the link prediction heuristics from a graph. The downside of WLNM
is that it truncates some neighbours to limit the sub-graph size to a user-defined size which
are may be informative for the prediction task.
5
4
International Journal of Data Mining & Knowledge Management Process (IJDKP), Vol.11, No.5, September 2021
major steps: (1) sub-graph extraction and node labelling, (2) node information matrix
construction, and (3) neural network training and evaluation. SEAL utilizes the available
information in the graph to improve the prediction performance. However, SEAL is limited
to be applied on homogeneous graphs though many real work graphs are heterogeneous
graphs. Moreover, the use of latent feature affects the computational time of SEAL.
where Ni is the set of neighbours for node ei. The output embedding of the node ei is
generated using the attention coefficients as in Equation 4.
→′
− Σ −
→
hi = αij W hj (4)
j∈Ni
GAT extends the single head concept to multi-head mechanism to learn more stable at-
tentions by averaging the coefficients over multi-head attentions. For link prediction, the
embedding of end nodes are feed into a fully connected NN.
6
6
International Journal of Data Mining & Knowledge Management Process (IJDKP), Vol.11, No.5, September 2021
takes the adjacency matrix (A), degree matrix(D), and a feature matrix(F) of graph as
inputs. The first layer generates a low-dimensional feature matrix as in Equation 5.
1 1
J J
X = GCN (F, A) = ReLU (A F W0) with A = J D− 2 AD− 2 (5)
where W0 is the weight matrix of first layer. The second layer generates the mean(µ) and
standard deviation(σ) of the Gaussian distribution as in Equation 6.
where Wµ, Wσ are weight matrices of second layer. The embeddings(Z) of nodes are
computed using parameterization trick as in Equation eq:zcal
The decoder is a generative model which is defined as the inner product between latent
variables to reconstructed adjacency matrix Â, which is defined as
 = σ(Z T Z) (8)
where σ(.) is the logistic sigmoid function. The weight matrices are learned by minimizing
the reconstruction loss capturing the similarity of A and Â.
The node embeddings(Z) are computed using the same way (Equation 10) as in VGAE.
The decoder remains same as the decoder in VGAE (8). The weight matrices Wµ and Wσ
are learned by minimizing the reconstruction loss as in VGAE.
3 EXPERIMENTAL DESIGN
Experimental data
We perform the comparative study of the above discussed similarity and embedding based
link prediction approaches in simple and undirected graphs from different domains. To
evaluate and describe the performance of the link prediction approaches, we choose ten
benchmark graphs from different areas: Ecoli [28], FB15K [29], NS [30], PB [31],Power [32],
Router [33], USAir [34], WN18 [35], YAGO3-10 [36], and Yeast [37]. FB1K, WN18 and
YAGO3-10 are popular knowledge graphs. These knowledge graphs consist of subject-
relationship type-object triples. However, as the studied approaches are applicable to
homogeneous graphs only. We simplify these knowledge graphs by overlooking the relation
names and considering links as undirected links.
7
7
International Journal of Data Mining & Knowledge Management Process (IJDKP), Vol.11, No.5, September 2021
The topological statistics of the graph datasets are summarized in Table 2. Based on
the number of nodes, these graphs are categorized into small/medium graphs with less or
equal 10,000 nodes and large graphs with more than 10,000 nodes.
8
7
International Journal of Data Mining & Knowledge Management Process (IJDKP), Vol.11, No.5, September 2021
9
7
International Journal of Data Mining & Knowledge Management Process (IJDKP), Vol.11, No.5, September 2021
4 EXPERIMENTAL RESULTS
The prediction approaches are evaluated in each of the five sets (train and test set) of
each graph and performance metrics (precision, AUC) are recorded. We measure the
precision in two different ways based on the top-L test links as described in Section 3.3. We
compute the threshold-based precision only for similarity-based approaches as embedding-
based approaches do learn the threshold. The auto-encoder based VGAE and LVGAE
approaches require the whole adjacency matrix of a graph on memory. Due to the memory
constaint, experiments with auto-encoder based VGAE and LVGAE approaches for large
scale graphs (WN18, YAGO3-10) failed. The maximum and minimum similarity scores
are computed from the top-L for each test set of each graph. Table 3 shows the results
in each of the seven small/medium and three large-size graphs. Each value of the table
is the mean over the five test sets. The standard deviation values of both metrics for all
approaches in all graphs are very small and they are not included in the table.
It can be clearly seen from Table 3 that the ranges of unnormalized similarity scores are
different for different similarity-based approaches and also different in different datasets
for the same similarity-based approach. Moreover, the minimum similarity scores are very
low (close to 0) in some datasets. These observations prove that in real-world applica-
tions, it is difficult to choose a threshold and to assess good precision for similarity-based
approaches.
From Table 1, the similarity-based approaches are mostly defined based on the common
neighbourhood. As expected, they show low precision (without defining threshold) and
AUC values in sparse graphs (low CC, low node degree) like Power, router and high preci-
sion for other well-connected graphs in Table 3. Exceptionally, PA shows better prediction
performance in sparse graphs as it considers individual node degree instead of common
neighbourhood for computing similarity score. The precision scores using the threshold-
based method drops drastically in most of the cases as many falsely predicted positive
links are identified (i.e. predicted links with very low scores). Surprisingly, HPI shows
competitive threshold-based precision value in NS dataset. No single similarity-based ap-
proach wins in all small/medium size graphs.
As expected, embedding-based approaches show very good precision and AUC scores
across all of the the small/medium size graphs compared to similarity-based approaches.
What about their comparative performances? No single approach wins in all datasets.
Node2Vec shows highest precision scores in some datasets though it is simpler than other
embedding-based approaches. The consideration of more distant neighbours in embed-
ding computation during random walk could be the most possible reason behind this
success. The use of latent information along with structural information in SEAL for
the datasets during prediction task likely explains the improvement of the metric AUC.
The best tuning of parameters could be the most possible reason behind the best balance
between the prediction metrics in GAT. The precision and AUC scores of auto-encoder
10
10
International Journal of Data Mining & Knowledge Management Process (IJDKP), Vol.11, No.5, September 2021
Table 3: AUC and Precision(Prec) values with Max Scores(Mx scr) and Min Scores (Mn
scr). Precision with * mark(Prec*) is computed based on threshold in top-L
links. Graph-wise highest metrics are indicated in bold fonts while approach-wise
highest metrics are shown in underline. – denotes the failed experiment.
App. Metric Ecoli NS PB Power Router USAir Yeast FB15K WN18 YAGO
3-10
Prec 0.90 0.87 0.86 0.17 0.07 0.92 0.83 0.77 0.13 0.15
Prec* 0.06 0.15 0.01 0.02 0.01 0.16 0.06 0.0002 0.0002 0.0018
AA Mx scr 32.84 5.83 33.41 3.04 5.60 16.69 23.71 418.60 57.32 24.44
Mn scr 2.86 1.14 0.58 0.00 0.00 2.70 0.00 0.12 0.00 0.00
AUC 0.93 0.94 0.92 0.58 0.54 0.94 0.91 0.82 0.56 0.48
Prec 0.78 0.69 0.83 0.49 0.41 0.85 0.79 0.79 0.63 0.83
PA Prec* 0.05 0.02 0.01 0.02 0.01 0.13 0.06 0.0003 0.0006 0.0006
Mx scr 65679 362.0 61052 53.0 2397 8298 10642 9881842 10637 2426939
Mn scr 3532 12.0 855.7 4.0 1.0 739.3 95.0 942.67 6.33 109.00
AUC 0.80 0.66 0.90 0.46 0.43 0.90 0.86 0.88 0.64 0.88
Prec 0.91 0.87 0.86 0.17 0.07 0.92 0.83 0.77 0.13 0.15
RA Prec* 0.03 0.15 0.01 0.03 0.01 0.10 0.07 0.0003 0.0002 0.0011
Mx scr 1.70 1.80 4.19 0.84 1.32 2.83 2.37 72.06 20.67 5.16
Mn scr 0.19 0.40 0.03 0.00 0.00 0.32 0.00 0.00 0.00 0.00
AUC 0.94 0.94 0.92 0.58 0.54 0.94 0.91 0.84 0.57 0.57
Prec 0.90 0.87 0.80 0.17 0.07 0.91 0.83 0.69 0.13 0.15
HPI Prec* 0.20 0.96 0.15 0.13 0.02 0.45 0.70 0.0959 0.0796 0.0476
Mx scr 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00
Mn scr 0.33 0.83 0.21 0.00 0.00 0.77 0.00 0.05 0.00 0.00
AUC 0.94 0.94 0.85 0.58 0.54 0.91 0.90 0.75 0.56 0.47
Prec 0.89 0.87 0.74 0.17 0.07 0.87 0.83 0.64 0.13 0.15
LLHN Prec* 0.001 0.13 0.001 0.03 0.003 0.03 0.01 0.0008 0.0046 0.0003
Mx scr 0.32 1.00 0.42 2.06 0.83 0.58 0.67 0.28 1.00 1.00
Mn scr 0.00 0.10 0.00 0.00 0.00 0.01 0.00 0.00 0.00 0.00
AUC 0.91 0.93 0.76 0.58 0.53 0.77 0.90 0.57 0.57 0.45
Prec 0.96 0.73 0.86 0.08 0.07 0.91 0.82 0.78 0.08 0.14
CCLP Prec* 0.06 0.21 0.01 0.01 0.01 0.18 0.06 0.0015 0.0006 0.0013
Mx scr 30.6 8.0 27.0 1.2 1.1 21.1 39.2 51.74 1.67 20.77
Mn scr 1.8 0.3 0.3 0.0 0.0 2.9 0.0 0.01 0.00 0.00
AUC 0.95 0.87 0.91 0.54 0.53 0.94 0.90 0.84 0.54 0.57
WLNM Prec 0.87 0.84 0.78 0.84 0.89 0.85 0.87 0.67 0.84 0.68
AUC 0.93 0.95 0.93 0.76 0.92 0.86 0.86 0.68 0.79 0.72
SEAL Prec 0.81 0.96 0.80 0.66 0.80 0.91 0.89 0.77 0.61 0.86
AUC 0.95 0.99 0.94 0.77 0.94 0.94 0.98 0.96 0.87 0.97
GAT Prec 0.84 0.93 0.84 0.72 0.81 0.88 0.91 0.85 0.74 0.84
AUC 0.85 0.90 0.86 0.70 0.79 0.87 0.89 0.87 0.79 0.83
Node2Vec Prec 0.91 0.97 0.91 0.86 0.80 0.81 0.85 0.79 0.83 0.82
AUC 0.90 0.96 0.90 0.82 0.75 0.85 0.94 0.88 0.79 0.80
VGAE Prec 0.75 0.91 0.86 0.82 0.80 0.78 0.75 0.81 – –
AUC 0.76 0.93 0.87 0.83 0.84 0.80 0.75 0.78 – –
LVGAE Prec 0.76 0.88 0.73 0.81 0.81 0.79 0.86 0.74 – –
AUC 0.77 0.86 0.73 0.86 0.83 0.80 0.83 0.68 – –
11
11
International Journal of Data Mining & Knowledge Management Process (IJDKP), Vol.11, No.5, September 2021
based approaches are behind the other embedding-based approaches. Improper tuning of
parameters could be the possible reason behind the performance degradation. Table 3
shows that embedding-based approaches provide high-performance metrics in all graphs
while similarity-based approaches perform well in some graphs (in terms of optimistic pre-
cision).
Considering the three large graphs (FB15K, WN18 and YAGO3-10), the prediction metrics
for similarity-based approaches are much lower than small/medium scale graphs, especially
in WN18 and YAGO3-10 graphs. Likewise the results in small/medium size graphs, the
precision scores of these approaches further drops drastically to less than 0.1 when applying
the threshold. Unsurprisingly, the prediction scores for embedding-based approaches in
large graphs are high as in small/medium scale graphs. The notable point in the prediction
metrics for large graphs is that Node2Vec is less competitive than other embedding-based
approaches in these large graphs.
We further compute the false positive rate, false negative rate and overall accuracy of the
approaches which are tabulated in Table 4. We compute the thresholds for similarity-
based approaches based on their top-L links scores which are still not optimal. Many of
existent and most of the test links have very low similarity scores. As a result, false posi-
tive rates of similarity-based approaches are very low and false negative rates is very high
for almost all graph datasets in Table 4. The overall accuracy scores of these approaches
are around 0.5. Exceptionally, HPI shows good accuracy in some datasets, especially in
NS dataset where the accuracy is more than 0.9. On the other hand, embedding-based ap-
proaches learn the threshold. The false positive rates of embedding-based approaches are
much higher than the similarity-based approaches, but the false negative rates are much
much lower than the similarity-based approaches. Among embedding-based approaches,
the auto-encoder based VGAE and LVGAE have higher false negative rates than other
embedding-based approaches. Overall, the accuracy scores of similarity-based approaches
stay around 0.5 with few exceptions. As expected, the accuracy scores of embedding based
approaches are much higher than the similarity-based approaches. With respect to accu-
racy, GAT shows highest/nearly highest scores in nearly half of the total graphs. And no
single embedding-based link prediction approach wins in all graphs.
Embedding-based link prediction approaches show better performance because they
learn heuristics from graphs. However, it is not clear which heuristic(s) are learned. We
want to take benefit from this study to get insight of such heuristics by comparing the
performances of similarity-based heuristics with performances of embedding-based ap-
proaches on the same datasets. In one hand, from Table. 1 and Table 3, AA, RA and
CCLP –which heuristically assign high weights to nodes with high degrees or cluster coef-
ficients – show better precision, accuracy on FB15K, PB, NS, USAir, and Yeast compared
to other graphs. GAT also shows better precision and accuracy these graphs than other
graphs. This may indicate that GAT learns similar heuristics as AA, RA and CCLP. In
the other hand, WLNM considers the role of each neighbour equally like HPI, LLHN, and
PA. WLNM, HPI, LLHN and PA show better performance scores on Power, Router, and
WN18 graphs, confirming that they are heuristically compatible. The auto-encoder based
VGAE and LVGAE approaches assign equal weights to all neighbours of a node when
aggregating neighbourhood information and show competitive performance on Power and
Router graphs. In addition, LVGAE aggregates hop-1 neighbours of a node using a simple
linear model whereas VGAE aggregates upto hop-2 neighbours of a node using 2-layer
GCN model. These indicate the possibility of learning HPI, LLHN and PA heuristics by
LVGAE and AA, RA, CCLP by VGAE approach.
12
12
International Journal of Data Mining & Knowledge Management Process (IJDKP), Vol.11, No.5, September 2021
Table 4: False positive rate(FPR), false negative rate (FNR) and accuracy(ACC) scores.
Graph-wise best metrics (lowest FPR, lowest FNR, highest ACC) are indicated
in bold fonts. – denotes the failed experiment.
App. Metric Ecoli NS PB Power Router USAir Yeast FB15K WN18 YAGO
3-10
FPR 0.0007 0.0001 0.0001 0.0003 0.0001 0.0028 0.0003 0.0001 0.0001 0.0001
AA FNR 0.962 0.898 0.991 0.977 0.996 0.932 0.956 0.998 0.997 0.998
ACC 0.519 0.551 0.504 0.512 0.502 0.533 0.522 0.499 0.499 0.499
FPR 0.0007 0.0007 0.0004 0.0018 0.0001 0.0019 0.0003 0.0001 0.0001 0.0001
PA FNR 0.967 0.942 0.993 0.988 0.996 0.952 0.956 0.997 0.999 0.999
ACC 0.516 0.529 0.503 0.505 0.502 0.523 0.522 0.499 0.499 0.498
FPR 0.0002 0.0001 0.0001 0.0003 0.0001 0.0000 0.0003 0.0001 0.0001 0.0001
RA FNR 0.984 0.854 0.995 0.977 0.994 0.973 0.958 0.998 0.999 0.999
ACC 0.508 0.573 0.502 0.511 0.503 0.514 0.521 0.499 0.499 0.498
FPR 0.0060 0.0058 0.0345 0.0012 0.0048 0.1175 0.0075 0.0067 0.001 0.0152
HPI FNR 0.753 0.160 0.863 0.877 0.974 0.531 0.412 0.904 0.920 0.953
ACC 0.620 0.917 0.551 0.561 0.510 0.676 0.790 0.543 0.538 0.514
FPR 0.0001 0.0001 0.0008 0.0003 0.0026 0.0114 0.0007 0.0013 0.0001 0.0002
LLHN FNR 0.999 0.826 0.999 0.985 0.998 0.988 0.994 0.999 0.995 0.999
ACC 0.501 0.587 0.499 0.507 0.500 0.500 0.503 0.498 0.501 0.498
FPR 0.0006 0.0001 0.0001 0.0001 0.0001 0.0019 0.0003 0.0001 0.0001 0.0001
CCLP FNR 0.959 0.927 0.991 0.992 0.992 0.919 0.952 0.998 0.999 0.989
ACC 0.520 0.536 0.504 0.504 0.504 0.539 0.524 0.494 0.499 0.499
FPR 0.1005 0.1642 0.2059 0.1310 0.1125 0.128 0.1334 0.3767 0.1466 0.4108
WLNM FNR 0.345 0.178 0.304 0.323 0.048 0.128 0.148 0.240 0.259 0.145
ACC 0.777 0.829 0.745 0.773 0.92 0.872 0.859 0.692 0.797 0.722
FPR 0.1774 0.0343 0.2214 0.2079 0.2323 0.0938 0.1044 0.2232 0.3492 0.1314
SEAL FNR 0.231 0.070 0.123 0.591 0.083 0.079 0.154 0.251 0.449 0.190
ACC 0.796 0.948 0.828 0.601 0.843 0.914 0.871 0.763 0.601 0.839
FPR 0.1828 0.0723 0.1909 0.2109 0.1328 0.1365 0.1025 0.1909 0.1848 0.1444
GAT FNR 0.073 0.069 0.032 0.461 0.428 0.046 0.015 0.032 0.3014 0.259
ACC 0.872 0.93 0.888 0.664 0.72 0.909 0.941 0.889 0.7569 0.7983
FPR 0.1457 0.0255 0.0826 0.1141 0.0541 0.1905 0.1454 0.1884 0.1594 0.1654
Node2Vec FNR 0.221 0.161 0.195 0.334 0.791 0.209 0.202 0.279 0.210 0.237
ACC 0.816 0.907 0.861 0.776 0.577 0.800 0.826 0.7663 0.815 0.7988
FPR 0.2455 0.0358 0.1306 0.0243 0.0352 0.2427 0.2399 0.1881 – –
VGAE FNR 0.257 0.661 0.316 0.891 0.861 0.151 0.289 0.304 – –
ACC 0.749 0.652 0.777 0.542 0.552 0.803 0.736 0.753 – –
FPR 0.1661 0.0971 0.2129 0.1478 0.1443 0.2464 0.1406 0.178 – –
LVGAE FNR 0.477 0.319 0.435 0.368 0.400 0.082 0.172 0.514 – –
ACC 0.678 0.792 0.676 0.742 0.728 0.836 0.844 0.651 – –
13
13
International Journal of Data Mining & Knowledge Management Process (IJDKP), Vol.11, No.5, September 2021
Table 5: Top-2 ranked similarity-based approaches with higher agreement with embedding-
based approach for test link decision. Numbers in () represent the agreement
percentages.
Graph WLNM SEAL GAT Node2Vec VGAE LGVAE
Ecoli HPI(69), LLHN(80), HPI(70), RA(70), RA(74), RA(72),
RA(69) RA(79) RA(69) LLHN(70) CCLP(74) HPI(72)
NS CCLP(65), AA(70), AA(61), AA(70), AA(65), PA(72),
AA(63) CCLP(68) PA(61) CCLP(68) RA(65) LLHN(72)
PB HPI(68), RA(68), LLHN(61), AA(68), RA(76), RA(71),
PA(64) PA(66) RA(59) RA(68) CCLP(75) AA(71)
Power HPI(63), PA(63), AA(67), PA(63), PA(54), PA(54),
LLHN(63) HPI(62) RA(67) RA(62) AA(51) LLHN(54)
Router PA(52), PA(66), CCLP(65), CCLP(69), PA(57), PA(52),
LLHN(47) LLHN(51) RA(65) AA(68) CCLP(54) HPI(52)
USAir AA(78), LLHN(90), CCLP(77), RA(90), RA(81), HPI(69),
CCLP(78) HPI(88) AA(75) LLHN(90) AA(81) RA(69)
Yeast CCLP(75), CCLP(70), CCLP(75), CCLP(70), AA(72), RA(69),
PA(74) AA(69) AA(71) AA(69) RAI(71) HPI(69)
FB15K RA(32), LLHN(30), HPI(28), HPI(26), HPI(42), HPI(38),
HPI(31) HPI(28) LLHN(27) AA(24) CCLP(39) LLHN(36)
WN18 PA(44), PA(40), PA(28), PA(36), – –
LLHN(42) HPI(32) AA(26) CCLP(31)
YAGO 3-10 PA(34), PA(44), PA(38), PA(42), – –
AA(26) AA(24) CCLP(32) RA(34)
14
14
International Journal of Data Mining & Knowledge Management Process (IJDKP), Vol.11, No.5, September 2021
Scalability: Many of the real-world graphs are huge in size. The scalability of embedding-
based approaches is achieved by sacrificing the completeness of graphs. In addition,
auto-encoder based approaches often failed for large scale graphs due to high mem-
ory requirement. A valuable future research could be developing a link prediction
approach in a distributed and parallel environment to make a good trade-off between
scalability and graph integrity.
Heterogenity: Most of the current embedding-based link prediction approaches are devel-
oped for homogeneous graphs. However, most of the real-world graphs are heteroge-
neous. A valuable future research could be designing a high-performing embedding-
based link prediction approach for heterogeneous.
6 CONCLUSION
In this paper, we study several link prediction approaches, looking for their performances
and connections. We focused on two categories of methods: similarity-based methods and
embedding-based learning methods. The studied approaches were evaluated on ten graph
datasets with different properties from various domains. The precision of similarity-based
approaches was computed in two different ways to highlight the difficulty of tuning the
threshold for deciding the link existence based on the similarity score. The experimental
results show the expected superiority of embedding-based approaches. Still, each of the
similarity-based approaches is competitive on graphs with specific properties. The possible
links between the handcrafted similarity-based approaches and current embedding-based
approaches were explored using (i) prediction performance comparison to get an idea
about the learned heuristics and (ii) agreement percentage on the diverse graphs. Our
15
15
International Journal of Data Mining & Knowledge Management Process (IJDKP), Vol.11, No.5, September 2021
References
[1] Z. Xu, C. Pu, and J. Yang, “Link prediction based on path entropy,” Physica A:
Statistical Mechanics and its Applications, vol. 456, pp. 294–301, 2016.
[2] Z. Shen, W.-X. Wang, Y. Fan, Z. Di, and Y.-C. Lai, “Reconstructing propagation
networks with natural diversity and identifying hidden sources,” Nature communi-
cations, vol. 5, no. 1, pp. 1–10, 2014.
[3] L. A. Adamic and E. Adar, “Friends and neighbors on the web,” Social networks,
vol. 25, no. 3, pp. 211–230, 2003.
[4] Y. Koren, R. Bell, and C. Volinsky, “Matrix factorization techniques for recom-
mender systems,” Computer, vol. 42, no. 8, pp. 30–37, 2009.
[5] M. Nickel, K. Murphy, V. Tresp, and E. Gabrilovich, “A review of relational machine
learning for knowledge graphs,” Proceedings of the IEEE, vol. 104, no. 1, pp. 11–33,
2015.
[6] I. A. Kovács, K. Luck, K. Spirohn, Y. Wang, C. Pollis, S. Schlabach, W. Bian, D.-K.
Kim, N. Kishore, T. Hao, et al., “Network-based prediction of protein interactions,”
Nature communications, vol. 10, no. 1, pp. 1–8, 2019.
[7] Z. Wu, S. Pan, F. Chen, G. Long, C. Zhang, and S. Y. Philip, “A comprehen-
sive survey on graph neural networks,” IEEE Transactions on Neural Networks and
Learning Systems, 2020.
[8] P. Cui, X. Wang, J. Pei, and W. Zhu, “A survey on network embedding,” IEEE
Transactions on Knowledge and Data Engineering, vol. 31, no. 5, pp. 833–852, 2018.
[9] V. Martı´nez, F. Berzal, and J.-C. Cubero, “A survey of link prediction in complex
networks,” ACM computing surveys (CSUR), vol. 49, no. 4, pp. 1–33, 2016.
[10] L. Lü and T. Zhou, “Link prediction in complex networks: A survey,” Physica A:
statistical mechanics and its applications, vol. 390, no. 6, pp. 1150–1170, 2011.
[11] H. Cai, V. W. Zheng, and K. C. C. Chang, “A comprehensive survey of graph em-
bedding: Problems, techniques, and applications,” IEEE Transactions on Knowledge
and Data Engineering, vol. 30, no. 9, pp. 1616–1637, 2018.
[12] M. K. Islam, S. Aridhi, and M. Smail-Tabbone, “Appraisal study of similarity-based
and embedding-based link prediction methods on graphs,” in Proceedings of the 10th
International Conference on Data Mining & Knowledge Management Process, 2021,
pp. 81–92.
[13] T. Zhou, L. Lü, and Y.-C. Zhang, “Predicting missing links via local information,”
The European Physical Journal B, vol. 71, no. 4, pp. 623–630, 2009.
[14] A.-L. Barabási and R. Albert, “Emergence of scaling in random networks,” science,
vol. 286, no. 5439, pp. 509–512, 1999.
[15] E. Ravasz, A. L. Somera, D. A. Mongru, Z. N. Oltvai, and A.-L. Barabási, “Hierar-
chical organization of modularity in metabolic networks,” science, vol. 297, no. 5586,
pp. 1551–1555, 2002.
16
15
International Journal of Data Mining & Knowledge Management Process (IJDKP), Vol.11, No.5, September 2021
17
15
International Journal of Data Mining & Knowledge Management Process (IJDKP), Vol.11, No.5, September 2021
[33] N. Spring, R. Mahajan, and D. Wetherall, “Measuring isp topologies with rocket-
fuel,” ACM SIGCOMM Computer Communication Review, vol. 32, no. 4, pp. 133–
145, 2002.
[34] M. S. Handcock, D. R. Hunter, C. T. Butts, S. M. Goodreau, and M. Morris,
“Statnet: An r package for the statistical modeling of social networks,” Web page
http://www. csde. washington. edu/statnet, 2003.
[35] A. Bordes, X. Glorot, J. Weston, and Y. Bengio, “A semantic matching energy
function for learning with multi-relational data,” Machine Learning, vol. 94, no. 2,
pp. 233–259, 2014.
[36] F. Mahdisoltani, J. Biega, and F. M. Suchanek, “Yago3: A knowledge base from
multilingual wikipedias,” in 7th Biennial Conference on Innovative Data Systems
Research, 2013.
[37] C. Von Mering, R. Krause, B. Snel, M. Cornell, S. G. Oliver, S. Fields, and P. Bork,
“Comparative assessment of large-scale datasets of protein- protein interactions,”
Nature, vol. 417, no. 6887, pp. 399–403, 2002.
[38] L. Pan, T. Zhou, L. Lü, and C. K. Hu, “Predicting missing links and identifying
spurious links via likelihood analysis,” Scientific Reports, vol. 6, no. 1, pp. 1–10,
2016.
[39] S. Tsugawa and H. Ohsaki, “Effectiveness of link prediction for face-to-face behav-
ioral networks,” Plos one, vol. 8, no. 12, e81727, 2013.
[40] Y. Yang, R. N. Lichtenwalter, and N. V. Chawla, “Evaluating link prediction meth-
ods,” Knowledge and Information Systems, vol. 45, no. 3, pp. 751–782, 2015.
[41] N. F. Liu, M. Gardner, Y. Belinkov, M. E. Peters, and N. A. Smith, “Linguistic
knowledge and transferability of contextual representations,” arXiv preprint arXiv:1903.08855,
2019.
[42] H. Azizpour, A. S. Razavian, J. Sullivan, A. Maki, and S. Carlsson, “Factors of
transferability for a generic convnet representation,” IEEE transactions on pattern
analysis and machine intelligence, vol. 38, no. 9, pp. 1790–1802, 2015.
[43] R. Levie, W. Huang, L. Bucci, M. M. Bronstein, and G. Kutyniok, “Transferability
of spectral graph convolutional neural networks,” arXiv preprint arXiv:1907.12972,
2019.
[44] J. Qiu, Q. Chen, Y. Dong, J. Zhang, H. Yang, M. Ding, K. Wang, and J. Tang, “Gcc:
Graph contrastive coding for graph neural network pre-training,” in Proceedings of
the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data
Mining, 2020, pp. 1150–1160.
[45] W. Hu, B. Liu, J. Gomes, M. Zitnik, P. Liang, V. Pande, and J. Leskovec, “Strategies
for pre-training graph neural networks,” arXiv preprint arXiv:1905.12265, 2019.
[46] L. Sun, Y. Dou, C. Yang, J. Wang, P. S. Yu, L. He, and B. Li, “Adversarial attack
and defense on graph data: A survey,” arXiv preprint arXiv:1812.10528, 2018.
18
15