Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

Influence Maximization Revisited: Efficient Sampling with Bound Tightened

Published: 18 August 2022 Publication History

Abstract

Given a social network G with n nodes and m edges, a positive integer k, and a cascade model C, the influence maximization (IM) problem asks for k nodes in G such that the expected number of nodes influenced by the k nodes under cascade model C is maximized. The state-of-the-art approximate solutions run in O(k(n+m)log n/ε2) expected time while returning a (1 - 1/e - ε) approximate solution with at least 1 - 1/n probability. A key phase of these IM algorithms is the random reverse reachable (RR) set generation, and this phase significantly affects the efficiency and scalability of the state-of-the-art IM algorithms.
In this article, we present a study on this key phase and propose an efficient random RR set generation algorithm under IC model. With the new algorithm, we show that the expected running time of existing IM algorithms under IC model can be improved to O(k ċ n log n ċ2), when for any node v, the total weight of its incoming edges is no larger than a constant. For the general IC model where the weights are skewed, we present a sampling algorithm SKIP. To the best of our knowledge, it is the first index-free algorithm that achieves the optimal time complexity of the sorted subset sampling problem.
Moreover, existing approximate IM algorithms suffer from scalability issues in high influence networks where the size of random RR sets is usually quite large. We tackle this challenging issue by reducing the average size of random RR sets without sacrificing the approximation guarantee. The proposed solution is orders of magnitude faster than states of the art as shown in our experiment.
Besides, we investigate the issues of forward propagation and derive its time complexity with our proposed subset sampling techniques. We also present a heuristic condition to indicate when the forward propagation approach should be utilized to estimate the expected influence of a given seed set.
Appendices

A Proof of Theorem 4

We introduce some notations and lemmas that are useful to prove Theorem 4. Denote \(\bar{\mu }\) as the total expected number of elements in \(T\) checked by SKIP to determine whether \(x_{j}\) is added to \(S\) (Lines 9– 10). Thus, SKIP takes \(O(1+\bar{\mu })\) time in expectation, where “1” is from the stopping step of while loop when it meets the condition \(j\gt h\) (Line 7).
Similarly, let \(\bar{\mu }(p_i,\ldots\,,p_{j})\) denote the expected number of elements checked when \(T=\lbrace x_i,\ldots ,x_j\rbrace\) , e.g., \(\bar{\mu }=\bar{\mu }(p_1,\ldots\,,p_{h})\) . Consider that SKIP performs on \(\lbrace x_i,\ldots ,x_j\rbrace\) and \(x_k\) is the first element checked. The probability for such a case is \((1-p_{i})^{k-i}p_{i}\) , as \(\Pr [X=k-i+1]\) follows the geometric distribution \(G(p_{i})\) . After checking \(x_k\) , SKIP performs on \(\lbrace x_{k+1},\ldots\,,x_j\rbrace\) . Thus, by Markov chain,
\begin{equation*} \bar{\mu }(p_i,\ldots ,p_{j})=\sum _{k=i}^{j} \left((1-p_i)^{k-i}p_i \cdot \left(1+\bar{\mu }(p_{k+1},\ldots ,p_{j})\right)\right). \end{equation*}
The following lemma shows the monotonicity of \(\bar{\mu }(p_i,\ldots\,,p_{j})\) .
Lemma 20.
\(\bar{\mu }(p_{i+1},\ldots\,,p_{j})\le \bar{\mu }(p_{i},\ldots\,,p_{j})\) .
Proof.
We prove the lemma by induction. When \(i=j-1\) , we have \(\bar{\mu }(p_{i+1})=p_{i+1}\) and \(\bar{\mu }(p_{i},p_{i+1})=p_i+(1-p_i)p_i+p_i p_{i+1}=(2+p_{i+1}-p_i)p_i\) , which indicates that \(\bar{\mu }(p_{i+1})\le \bar{\mu }(p_{i},p_{i+1})\) . Assume that for any \(i\ge i^\ast\) , it holds that
\begin{equation*} \bar{\mu }(p_{i+1},\ldots\,,p_{j})\le \bar{\mu }(p_{i},\ldots\,,p_{j}). \end{equation*}
Now consider \(i=i^\ast -1\) . Similar to () we have
\begin{equation*} \bar{\mu }(p_{i+1},\ldots\,,p_{j})=\sum _{k=i+1}^{j} \left((1-p_{i+1})^{k-i-1}p_{i+1} \cdot \left(1+\bar{\mu }(p_{k+1},\ldots\,,p_{j})\right)\right). \end{equation*}
For any \(\ell =i,\ldots\,,j\) , define
\begin{equation*} \Delta _\ell :=\sum _{k=i}^{\ell } (1-p_i)^{k-i}p_i -\sum _{k=i+1}^{\ell } (1-p_{i+1})^{k-i-1}p_{i+1} =(1-p_{i+1})^{\ell -i}-(1-p_i)^{\ell -i+1}. \end{equation*}
It is easy to verify that \(\Delta _\ell \ge 0\) for any \(\ell = i, \ldots\,, j\) owing to the fact that \(\lbrace p_i, \ldots\,, p_j\rbrace\) are in non-ascending order. Besides, the following conclusion holds by definition, and will be used later,
\begin{equation} \Delta _{\ell +1} + p_{i+1}(1-p_{i+1})^{l-i} = \Delta _{\ell } + p_i(1-p_i)^{l-i+1}. \end{equation}
(20)
Then, we have
\begin{align*} &\bar{\mu }(p_{i},\ldots\,,p_{j})=\sum _{k=i}^{j} \left((1-p_i)^{k-i}p_i \cdot \left(1+\bar{\mu }(p_{k+1},\ldots\,,p_{j})\right)\right) \\ &=p_i\left(1+\bar{\mu }(p_{i+1},\ldots\,,p_{j})\right)+\sum _{k=i+1}^{j} \left((1-p_{i})^{k-i}p_{i} \cdot \left(1+\bar{\mu }(p_{k+1},\ldots\,,p_{j})\right)\right)\\ &\ge \Delta _{i+1}\left(1+\bar{\mu }(p_{i+2},\ldots\,,p_{j})\right)+p_{i+1}\left(1+\bar{\mu }(p_{i+2},\ldots\,,p_{j})\right)\\ &\quad {+}\,\sum _{k=i+2}^{j} \left((1-p_{i})^{k-i}p_{i} \cdot \left(1+\bar{\mu }(p_{k+1},\ldots\,,p_{j})\right)\right). \end{align*}
The inequality is due to \(\bar{\mu }(p_{i+1},\ldots\,,p_{j})\ge \bar{\mu }(p_{i+2},\ldots\,,p_{j})\) and \(\Delta _{i+1}+p_{i+1}=p_{i}+(1-p_i)p_i\) by Equation (20).
Recursively, we have
\begin{align*} \bar{\mu }(p_{i},\ldots\,,p_{j}) &\ge \Delta _{i+2}\left(1+\bar{\mu }(p_{i+3},\ldots\,,p_{j})\right) {+}\sum _{k=i+1}^{i+2}\left((1-p_{i+1})^{k-i-1}p_{i+1} \cdot \left(1+\bar{\mu }(p_{k+1},\ldots\,,p_{j})\right)\right)\\ &\quad +\sum _{k=i+3}^{j} \left((1-p_{i})^{k-i}p_{i} \cdot \left(1+\bar{\mu }(p_{k+1},\ldots\,,p_{j})\right)\right)\\ &\ge \cdots \ge \Delta _{j}+\bar{\mu }(p_{i+1},\ldots\,,p_{j}). \end{align*}
Note that \(\Delta _j\ge 0\) as \(p_i\ge p_{i+1}\) , which immediately concludes the lemma.□
Lemma 21.
For any \(p_k\le p_k^\prime\) , \(\bar{\mu }(p_{i},\ldots\,,p_k,\ldots\,,p_{j})\le \bar{\mu }(p_{i},\ldots\,,p_k^\prime ,\ldots\,,p_{j})\) .
Proof.
We prove the lemma by induction. When \(i=j=k\) , it obviously holds that \(\bar{\mu }(p_k)=p_k\le p_k^\prime =\bar{\mu }(p_k^\prime)\) . Assume that \(\bar{\mu }(p_{i},\ldots\,,p_k,\ldots\,,p_{j})\le \bar{\mu }(p_{i},\ldots\,,p_k^\prime ,\ldots\,,p_{j})\) that for any \(i\ge i^\ast\) . Now consider the following two cases when \(i=i^\ast -1\) .
Case (i) \(k\gt i\) . With assumption that for any \(\ell +1\ge i+1\ge i^\ast\) , \(\bar{\mu }(p_{\ell +1},\ldots\,,p_k,\ldots\,,p_{j})\le \bar{\mu }(p_{\ell +1},\ldots\,,p_k^\prime ,\ldots\,,p_{j})\) , then we have:
\begin{align*} \bar{\mu }(p_{i},\ldots\,,p_k,\ldots\,,p_{j}) &=\sum _{\ell =i}^{j} \left((1-p_i)^{\ell -i}p_i \cdot \left(1+\bar{\mu }(p_{\ell +1},\ldots\,,p_k,\ldots\,,p_{j})\right)\right)\\ &\le \sum _{\ell =i}^{j} \left((1-p_i)^{\ell -i}p_i \cdot \left(1+\bar{\mu }(p_{\ell +1},\ldots\,,p_k^\prime ,\ldots\,,p_{j})\right)\right)=\bar{\mu }(p_{i},\ldots\,,p_k^\prime ,\ldots\,,p_{j}). \end{align*}
Case (ii) \(k=i\) . Let \(\Gamma _\ell :=\sum _{i=k}^{\ell } (1-p_k^\prime)^{i-k}p_k^\prime -\sum _{i=k}^{\ell }(1-p_k)^{i-k}p_k\) , which implies that \(\Gamma _\ell =(1-p_k)^{\ell -k+1}-(1-p_k^\prime)^{\ell -k+1}\ge 0\) for any \(\ell =k,\ldots\,,j\) , since \(p_k\le p_k^\prime\) . Similar with (20), we have
\begin{equation} -\Gamma _{\ell +1} + p_k^\prime (1-p_k^\prime)^{\ell -k +1} = - \Gamma _{\ell } + p_k(1-p_k)^{\ell -k+1}. \end{equation}
(21)
Then, we can get that
\begin{align*} & \bar{\mu }(p_k,\ldots\,,p_{j}) = \sum _{\ell =k}^{j} \left((1-p_k)^{\ell -k}p_k \cdot \left(1+\bar{\mu }(p_{\ell +1},\ldots\,,p_{j})\right)\right)\\ &= -\Gamma _k\left(1+\bar{\mu }(p_{k+1},\ldots\,,p_{j})\right)+p_k^\prime \left(1+\bar{\mu }(p_{k+1},\ldots\,,p_{j})\right)\\ &\quad +\,\sum _{\ell =k+1}^{j} \left((1-p_k)^{\ell -k}p_k \cdot \left(1+\bar{\mu }(p_{\ell +1},\ldots\,,p_{j})\right)\right)\\ &\le -\Gamma _{k+1}\left(1+\bar{\mu }(p_{k+2},\ldots\,,p_{j})\right) {+}\sum _{\ell =k}^{k+1} \left((1-p_k^\prime)^{\ell -k}p_k^\prime \cdot \left(1+\bar{\mu }(p_{\ell +1},\ldots\,,p_{j})\right)\right) \\ &\quad +\sum _{\ell =k+2}^{j} \left((1-p_k)^{\ell -k}p_k \cdot \left(1+\bar{\mu }(p_{\ell +1},\ldots\,,p_{j})\right)\right)\\ &\le \cdots \le -\Gamma _{j}+\bar{\mu }(p_k^\prime ,\ldots\,,p_{j}) \le \bar{\mu }(p_k^\prime ,\ldots\,,p_{j}), \end{align*}
where the inequality is due to Lemma 20 and (21). Thus, Lemma 21 follows.□
Given any \(\beta =2,\ldots\,,n\) , we divide \(V\) into \(L(\beta)+1\) buckets such that bucket \(B_k=\lbrace x_i:\beta ^{k}\le i\lt \beta ^{k+1}\rbrace\) with \(k=0,1,\ldots\,,L(\beta)\) , where \(L(\beta)=\lfloor \log _\beta n\rfloor\) . For \(x_i\in B_k\) , let \(\hat{p}_i:=p_{\beta ^k}\) which is an upper bound on \(p_i\) . The following lemma establishes an upper bound on \(\bar{\mu }\) .
Lemma 22.
Given \(\beta =2,\ldots\,,n\) , \(\bar{\mu }\le \hat{\mu }(\beta)+L(\beta) + 1\) , where \(\hat{\mu }(\beta)=\sum _{i=1}^{n}\hat{p}_i\) .
Proof.
Define \(\bar{\mu }^\ast :=\bar{\mu }(\hat{p}_1,\ldots\,,\hat{p}_n)\) . By Lemma 21, we have \(\bar{\mu }\le \bar{\mu }^\ast\) . We just need to show that \(\bar{\mu }^\ast \le \hat{\mu }(\beta)+L(\beta)+1\) . By Lemma 20, we have
\begin{align*} \bar{\mu }^\ast &=\sum _{k=1}^{n} \left((1-\hat{p}_1)^{k-1} \cdot \left(1+\bar{\mu }(\hat{p}_{k+1},\ldots\,,\hat{p}_{n})\right)\right)\\ &\le \sum _{k=1}^{\beta -1} \left((1-\hat{p}_1)^{k-1}\hat{p}_1 \cdot \left(1+\bar{\mu }(\hat{p}_{k+1},\ldots\,,\hat{p}_{n})\right)\right) {+}\sum _{k=\beta }^{n} \left((1-\hat{p}_1)^{k-1}\hat{p}_1 \cdot \left(1+\bar{\mu }(\hat{p}_{\beta },\ldots\,,\hat{p}_{n})\right)\right)\\ &\le \sum _{k=1}^{\beta -1} \left((1-\hat{p}_1)^{k-1}\hat{p}_1 \cdot \left(1+\bar{\mu }(\hat{p}_{k+1},\ldots\,,\hat{p}_{n})\right)\right) {+}(1-\hat{p}_1)^{\beta -1}\cdot \left(1+\bar{\mu }(\hat{p}_{\beta },\ldots\,,\hat{p}_{n})\right). \end{align*}
For \(\beta = 2\) , it is straightforward to verify that,
\begin{align*} \mu ^\ast &\le \hat{p}_1 \left(1 + \bar{\mu }(\hat{p}_{\beta },\ldots\,,\hat{p}_{n})\right) + (1- \hat{p}_1)\left(1+\bar{\mu }(\hat{p}_{\beta },\ldots\,,\hat{p}_{n})\right)\\ &= 1+\bar{\mu }(\hat{p}_{\beta },\ldots\,,\hat{p}_{n}) \le \hat{p}_1 + 1+\bar{\mu }(\hat{p}_{\beta },\ldots\,,\hat{p}_{n}). \end{align*}
For \(\beta \gt 2\) , we have
\begin{align*} \bar{\mu }(\hat{p}_{2},\ldots\,,\hat{p}_{n}) &\le \sum _{k=2}^{\beta -1} \left((1-\hat{p}_2)^{k-2}\hat{p}_2 \cdot \left(1+\bar{\mu }(\hat{p}_{k+1},\ldots\,,\hat{p}_{n})\right)\right) {+}(1-\hat{p}_2)^{\beta -2}\cdot \left(1+\bar{\mu }(\hat{p}_{\beta },\ldots\,,\hat{p}_{n})\right). \end{align*}
Note that \(\hat{p}_1=\hat{p}_2\) when \(\beta \gt 2\) . Thus, \((1-\hat{p}_1)^{k-1}\hat{p}_1 +\hat{p}_1(1-\hat{p}_2)^{k-2}\hat{p}_2=(1-\hat{p}_1)^{k-2} \hat{p}_1.\) As a result,
\begin{align*} \bar{\mu }^\ast &\le \hat{p}_1 + \sum _{k=2}^{\beta -1} \left(\hat{p}_1 (1 - \hat{p}_2)^{k-2} \hat{p}_2 + (1 - \hat{p}_1)^{k-1} \hat{p}_1) \right)\left(1+\bar{\mu }(\hat{p}_{k+1},\ldots\,,\hat{p}_{n}) \right) \\ &\quad + \left(\hat{p}_1 (1 - p_2)^{\beta - 2} + (1 - \hat{p}_1)^{\beta - 1}\right) \left(1+\bar{\mu }(\hat{p}_{\beta },\ldots\,,\hat{p}_{n})\right)\\ &\le \hat{p}_1+\sum _{k=2}^{\beta -1} \left((1-\hat{p}_1)^{k-2}\hat{p}_1 \cdot \left(1+\bar{\mu }(\hat{p}_{k+1},\ldots\,,\hat{p}_{n})\right)\right){+}(1-\hat{p}_1)^{\beta -2}\cdot \left(1+\bar{\mu }(\hat{p}_{\beta },\ldots\,,\hat{p}_{n})\right)\\ &\le \cdots \le (\beta -2)\hat{p}_1+\hat{p}_1 \cdot \left(1+\bar{\mu }(\hat{p}_{\beta },\ldots\,,\hat{p}_{n})\right)+(1-\hat{p}_1)\cdot \left(1+\bar{\mu }(\hat{p}_{\beta },\ldots\,,\hat{p}_{n})\right)\\ &\le (\beta -1)\hat{p}_1+1+\bar{\mu }(\hat{p}_{\beta },\ldots\,,\hat{p}_{n}). \end{align*}
Therefore for \(\beta = 2, \ldots\,, n\) , we recursively have
\begin{align*} \bar{\mu }^\ast &\le (\beta -1)\hat{p}_1+1+\bar{\mu }(\hat{p}_{\beta },\ldots\,,\hat{p}_{n}) \le (\beta -1)\hat{p}_1+(\beta ^2-\beta)\hat{p}_{\beta }+2+\bar{\mu }(\hat{p}_{\beta ^2},\ldots\,,\hat{p}_{n})\\ &\le \cdots \le \sum _{k=0}^{L(\beta)-1}(\beta ^{k+1}-\beta ^{k})\hat{p}_{\beta ^k}+L(\beta)+\bar{\mu }(\hat{p}_{\beta ^{L(\beta)}},\ldots\,,\hat{p}_{n}). \end{align*}
Meanwhile, SKIP performs sampling with standard geometric distribution \(G(\hat{p}_{\beta ^{L(\beta)}})\) on \(\lbrace \hat{p}_{\beta ^{L(\beta)}},\ldots\,,\hat{p}_{n}\rbrace\) , which indicates that \(\bar{\mu }(\hat{p}_{\beta ^{L(\beta)}},\ldots\,,\hat{p}_{n})=(n-\beta ^{L(\beta)}+1)\hat{p}_{\beta ^{L(\beta)}} +1\) . Therefore, \(\bar{\mu }^\ast \le \hat{\mu }(\beta)+L(\beta) + 1\) .□
Now, we are ready to prove our main result.
Proof of Theorem 4
Recall that SKIP takes an expected time of \(O(1+\bar{\mu })\) . By Lemma 22, we know that \(\bar{\mu }\le \min _{\beta \in \lbrace 2,\ldots\,,n\rbrace }\lbrace \hat{\mu }(\beta)+L(\beta) + 1\rbrace\) . In addition, according to the definition of \(\hat{p}_i\) under a given \(\beta\) , it is easy to verify that \(\hat{p}_i\le p_{\lceil i/\beta \rceil }\) . Thus,
\begin{equation*} \hat{\mu }(\beta)=\sum _{i=1}^{n}\hat{p}_i\le \sum _{i=1}^{n}p_{\lceil i/\beta \rceil }\le \beta \sum _{i=1}^{n}p_i=\beta \mu . \end{equation*}
Therefore, \(\bar{\mu }\le \min _{\beta \in \lbrace 2,\ldots\,,n\rbrace }\lbrace \beta \mu +\log _\beta n + 1\rbrace\) .
When \(\mu \ge {(\log n)}/{2}\) , \(\bar{\mu }\le 2\mu +\log _2 n\le 6\mu\) by setting \(\beta =2\) . Thus, Theorem 4 holds, since \(O(1+\bar{\mu })=O(\mu)\) .
Next, we consider \(\mu \lt {(\log n)}/{2}\) . Define
\begin{equation*} \gamma :=\frac{({\log n})/{\mu }}{\log \left(({\log n})/{\mu }\right)}\quad \text{and}\quad \beta ^\ast :=\lceil \gamma \rceil . \end{equation*}
Thus, \(\beta ^\ast \mu =O(\frac{{\log n}}{\log (({\log n})/{\mu })})\) and \(\log _{\beta ^\ast } n=O(\frac{{\log n}}{\log (({\log n})/{\mu })})\) . Therefore,
\begin{equation*} O(1+\bar{\mu })=O\left(1+\frac{{\log n}}{\log (({\log n})/{\mu })}\right). \end{equation*}
This completes the proof.□

B Variant of Greedy Algorithm

In this appendix, we present another revised greedy algorithm, which is different from Greedy-Degree in Algorithm 9. For ease of explanation, we name it Greedy-Cost.
Recap that in the second phase of HIST, we can stop the RR set generation process as soon as we hit any sentinel node, thus reducing the average size of the RR sets. Suppose we have a function \(C_\mathcal {R}(S)\) which represents the amount of cost reduction on the collection \(\mathcal {R}\) of RR sets when \(S\) is selected as the sentinel set. Then we can design the Greedy-Cost algorithm by replacing Line 3 in Algorithm 1, that is,
\begin{equation*} v \leftarrow \operatorname{arg\,max}_{u\in V}(\Lambda _\mathcal {R}(S^*_{k}\cup \lbrace u\rbrace))-\Lambda _{\mathcal {R}}(S^*_k), \end{equation*}
with the following statements:
\begin{align*} &\mathcal {M} \leftarrow \operatorname{arg\,max}_{u\in V}(\Lambda _\mathcal {R}(S^*_{k}\cup \lbrace u\rbrace))-\Lambda _{\mathcal {R}}(S^*_k) \\ &v \leftarrow \operatorname{arg\,max}_{v^{\prime } \in \mathcal {M}} C_\mathcal {R}(S^*_{k}\cup \lbrace v^{\prime }\rbrace) - C_\mathcal {R}(S^*_k), \end{align*}
where \(\mathcal {M}\) is the set of nodes which have the largest marginal coverage. Obviously if \(|\mathcal {M}| = 1\) , we have only one candidate, and it must be selected as the sentinel node.
In the following, we present how to define the cost funtion \(C_\mathcal {R}(\cdot)\) . Recap that when generating an RR set \(R\) , each sampled node is added into \(R\) one by one (please see Algorithm 2). Thus, we say the index of \(u\) is \(i\) if it is the \(i\) th node added into \(R\) , where \(i = 0, 1, 2, \ldots , |R|-1\) . Let \(l(u, R)\) be the function which returns the index of \(u\) in \(R\) . Let \(l(u, R) = |R|\) if \(u \not\in R\) . Due to the existence of the sentinel set \(S\) , the process of generating \(R\) can be stopped immediately when it reaches the sentinel node \(u^* \in S\) with the minimum index,
\begin{align*} u^* = \operatorname{arg\,min}_{u^{\prime } \in S} l(u^{\prime }, R). \end{align*}
Therefore, we define the cost reduction function on an RR set \(R\) caused by the sentinel set \(S\) as
\begin{equation*} C(R, S) = |R| - l(u^*, R). \end{equation*}
It is easy to realize that if no node of \(S\) is hit by \(R\) , the cost reduction \(C(R, S) = 0\) due to \(l(u, R) = |R|\) for each \(u \in S\) . It implies that it can not save any sampling cost when generating \(R\) . By summing up all the cost reduction in \(\mathcal {R}\) , the function \(C_\mathcal {R}(S)\) is defined as
\begin{equation*} C_\mathcal {R}(S) = \sum _{R \in \mathcal {R}} C(R, S). \end{equation*}

References

[1]
2013. KONECT Datasets. Retrieved June 2019 from http://konect.uni-koblenz.de/. (2013).
[2]
2014. PMC code. Retrieved December 2021 from https://github.com/todo314/pruned-monte-carlo. (2014).
[3]
2014. SNAP Datasets. Retrieved June 2019 from http://snap.stanford.edu/data. (2014).
[4]
2015. IMM code. Retrieved June 2019 from https://sourceforge.net/projects/im-imm/. (2015).
[5]
2017. OPIM-C code. Retrieved June 2019 from https://github.com/tangj90/OPIM. (2017).
[6]
2017. SSA code. Retrieved June 2019 from https://github.com/hungnt55/Stop-and-Stare. (2017).
[7]
Akhil Arora, Sainyam Galhotra, and Sayan Ranu. 2017. Debunking the myths of influence maximization: An in-depth benchmarking study. In Proceedings of the SIGMOD. 651–666.
[8]
Song Bian, Qintian Guo, Sibo Wang, and Jeffrey Xu Yu. 2020. Efficient algorithms for budgeted influence maximization on massive social networks. Proceedings of the VLDB Endowment 13, 9 (2020), 1498–1510.
[9]
Christian Borgs, Michael Brautbar, Jennifer T. Chayes, and Brendan Lucier. 2014. Maximizing social influence in nearly optimal time. In Proceedings of the SODA. 946–957.
[10]
Stephen Boyd, Stephen P. Boyd, and Lieven Vandenberghe. 2004. Convex Optimization. Cambridge University Press.
[11]
Karl Bringmann and Konstantinos Panagiotou. 2017. Efficient sampling methods for discrete distributions. Algorithmica 79, 2 (2017), 484–508.
[12]
Ceren Budak, Divyakant Agrawal, and Amr El Abbadi. 2011. Limiting the spread of misinformation in social networks. In Proceedings of the WWW. 665–674.
[13]
Shuo Chen, Ju Fan, Guoliang Li, Jianhua Feng, Kian-Lee Tan, and Jinhui Tang. 2015. Online topic-aware influence maximization. PVLDB 8, 6 (2015), 666–677.
[14]
Wei Chen, Chi Wang, and Yajun Wang. 2010. Scalable influence maximization for prevalent viral marketing in large-scale social networks. In Proceedings of the SIGKDD. 1029–1038.
[15]
Wei Chen, Yajun Wang, and Siyu Yang. 2009. Efficient influence maximization in social networks. In Proceedings of the SIGKDD. 199–208.
[16]
Suqi Cheng, Huawei Shen, Junming Huang, Wei Chen, and Xueqi Cheng. 2014. IMRank: Influence maximization via finding self-consistent ranking. In Proceedings of the SIGIR. 475–484.
[17]
Fan R. K. Chung and Lincoln Lu. 2006. Survey: Concentration inequalities and martingale inequalities: A survey. Internet Mathematics 3, 1 (2006), 79–127.
[18]
Edith Cohen, Daniel Delling, Thomas Pajor, and Renato F. Werneck. 2014. Sketch-based influence maximization and computation: Scaling up with guarantees. In Proceedings of the CIKM. 629–638.
[19]
Paul Dagum, Richard M. Karp, Michael Luby, and Sheldon M. Ross. 1995. An optimal algorithm for monte carlo estimation (Extended Abstract). In Proceedings of the FOCS. 142–149.
[20]
Sainyam Galhotra, Akhil Arora, and Shourya Roy. 2016. Holistic influence maximization: Combining scalability and efficiency with opinion-aware models. In Proceedings of the SIGMOD. 743–758.
[21]
Manuel Gomez-Rodriguez, David Balduzzi, and Bernhard Schölkopf. 2011. Uncovering the temporal dynamics of diffusion networks. In Proceedings of the ICML. 561–568.
[22]
Amit Goyal, Francesco Bonchi, and Laks V. S. Lakshmanan. 2010. Learning influence probabilities in social networks. In Proceedings of the WSDM. 241–250.
[23]
Amit Goyal, Francesco Bonchi, and Laks V. S. Lakshmanan. 2011. A data-based approach to social influence maximization. PVLDB 5, 1 (2011), 73–84.
[24]
Amit Goyal, Wei Lu, and Laks V. S. Lakshmanan. 2011. CELF++: Optimizing the greedy algorithm for influence maximization in social networks. In Proceedings of the WWW. 47–48.
[25]
Amit Goyal, Wei Lu, and Laks V. S. Lakshmanan. 2011. SIMPATH: An efficient algorithm for influence maximization under the linear threshold model. In Proceedings of the ICDM. 211–220.
[26]
Qintian Guo, Sibo Wang, Zhewei Wei, and Ming Chen. 2020. Influence maximization revisited: Efficient reverse reachable set generation with bound tightened. In Proceedings of the SIGMOD. ACM, 2167–2181.
[27]
Kai Han, Keke Huang, Xiaokui Xiao, Jing Tang, Aixin Sun, and Xueyan Tang. 2018. Efficient algorithms for adaptive influence maximization. PVLDB 11, 9 (2018), 1029–1040.
[28]
Keke Huang, Sibo Wang, Glenn S. Bevilacqua, Xiaokui Xiao, and Laks V. S. Lakshmanan. 2017. Revisiting the stop-and-stare algorithms for influence maximization. PVLDB 10, 9 (2017), 913–924.
[29]
Kyomin Jung, Wooram Heo, and Wei Chen. 2012. IRIE: Scalable and robust influence maximization in social networks. In Proceedings of the ICDM. 918–923.
[30]
David Kempe, Jon M. Kleinberg, and Éva Tardos. 2003. Maximizing the spread of influence through a social network. In Proceedings of the SIGKDD. 137–146.
[31]
Donald Ervin Knuth. 1997. The Art of Computer Programming. Vol. 3. Addison-Wesley.
[32]
Andreas Krause and Daniel Golovin. 2014. Submodular function maximization. In Proceedings of the Tractability: Practical Approaches to Hard Problems. Cambridge University, 71–104.
[33]
Siyu Lei, Silviu Maniu, Luyi Mo, Reynold Cheng, and Pierre Senellart. 2015. Online influence maximization. In Proceedings of the SIGKDD. 645–654.
[34]
Yuchen Li, Dongxiang Zhang, and Kian-Lee Tan. 2015. Real-time targeted influence maximization for online advertisements. PVLDB 8, 10 (2015), 1070–1081.
[35]
Bo Liu, Gao Cong, Dong Xu, and Yifeng Zeng. 2012. Time constrained influence maximization in social networks. In Proceedings of the ICDM. 439–448.
[36]
Wei Lu, Wei Chen, and Laks V. S. Lakshmanan. 2015. From competition to complementarity: Comparative influence diffusion and maximization. PVLDB 9, 2 (2015), 60–71.
[37]
Brendan Lucier, Joel Oren, and Yaron Singer. 2015. Influence at scale: Distributed computation of complex contagion in networks. In Proceedings of the SIGKDD. 735–744.
[38]
Hung T. Nguyen, Thang N. Dinh, and My T. Thai. 2016. Cost-aware targeted viral marketing in billion-scale networks. In Proceedings of the INFOCOM. 1–9.
[39]
Hung T. Nguyen, Thang N. Dinh, and My T. Thai. 2018. Revisiting of “Revisiting the Stop-and-Stare Algorithms for Influence Maximization”. In Proceedings of the CSoNet. 273–285.
[40]
Hung T. Nguyen, Tri P. Nguyen, NhatHai Phan, and Thang N. Dinh. 2017. Importance sketching of influence dynamics in billion-scale networks. In Proceedings of the ICDM. 337–346.
[41]
Hung T. Nguyen, Tri P. Nguyen, Tam N. Vu, and Thang N. Dinh. 2017. Outward influence and cascade size estimation in billion-scale networks. Proceedings of the ACM on Measurement and Analysis of Computing Systems 1, 1 (2017), 20:1–20:30.
[42]
Hung T. Nguyen, My T. Thai, and Thang N. Dinh. 2016. Stop-and-stare: Optimal sampling algorithms for viral marketing in billion-scale networks. In Proceedings of the SIGMOD. 695–710.
[43]
Naoto Ohsaka, Takuya Akiba, Yuichi Yoshida, and Ken-ichi Kawarabayashi. 2014. Fast and accurate influence maximization on large networks with pruned monte-carlo simulations. In Proceedings of the AAAI. 138–144.
[44]
Grant Schoenebeck and Biaoshuai Tao. 2020. Influence maximization on undirected graphs: Toward closing the (1-1/e) gap. ACM Transactions on Economics and Computation 8, 4 (2020), 22:1–22:36.
[45]
Jing Tang, Keke Huang, Xiaokui Xiao, Laks V. S. Lakshmanan, Xueyan Tang, Aixin Sun, and Andrew Lim. 2019. Efficient approximation algorithms for adaptive seed minimization. In Proceedings of the SIGMOD. 1096–1113.
[46]
Jing Tang, Xueyan Tang, Xiaokui Xiao, and Junsong Yuan. 2018. Online processing algorithms for influence maximization. In Proceedings of the SIGMOD. 991–1005.
[47]
Youze Tang, Yanchen Shi, and Xiaokui Xiao. 2015. Influence maximization in near-linear time: A martingale approach. In Proceedings of the SIGMOD. 1539–1554.
[48]
Youze Tang, Xiaokui Xiao, and Yanchen Shi. 2014. Influence maximization: Near-optimal time complexity meets practical efficiency. In Proceedings of the SIGMOD. 75–86.
[49]
Rajan Udwani. 2018. Multi-objective maximization of monotone submodular functions with cardinality constraint. In Proceedings of the NeurIPS. 9513–9524.
[50]
Alastair J. Walker. 1977. An efficient method for generating discrete random variables with general distributions. ACM Transactions on Mathematical Software 3, 3 (1977), 253–256.
[51]
Yanhao Wang, Qi Fan, Yuchen Li, and Kian-Lee Tan. 2017. Real-time influence maximization on dynamic social streams. PVLDB 10, 7 (2017), 805–816.

Cited By

View all
  • (2024)Link Recommendation to Augment Influence Diffusion with Provable GuaranteesProceedings of the ACM Web Conference 202410.1145/3589334.3645521(2509-2518)Online publication date: 13-May-2024
  • (2024)Generalized hop‐based approaches for identifying influential nodes in social networksExpert Systems10.1111/exsy.13649Online publication date: 4-Jun-2024
  • (2024)Conformity-aware adoption maximization in competitive social networksNeurocomputing10.1016/j.neucom.2023.127224573(127224)Online publication date: Mar-2024
  • Show More Cited By

Index Terms

  1. Influence Maximization Revisited: Efficient Sampling with Bound Tightened

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Transactions on Database Systems
    ACM Transactions on Database Systems  Volume 47, Issue 3
    September 2022
    173 pages
    ISSN:0362-5915
    EISSN:1557-4644
    DOI:10.1145/3544001
    Issue’s Table of Contents

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 18 August 2022
    Online AM: 19 May 2022
    Accepted: 01 April 2022
    Revised: 01 February 2022
    Received: 01 June 2021
    Published in TODS Volume 47, Issue 3

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Influence maximization
    2. sampling

    Qualifiers

    • Research-article
    • Refereed

    Funding Sources

    • Hong Kong RGC ECS
    • Hong Kong RGC CRF
    • Hong Kong ITC ITF
    • CUHK Direct Grant
    • NSFC of China
    • National Natural Science Foundation of China
    • Beijing Natural Science Foundation
    • PCL
    • HKUST(GZ)

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)299
    • Downloads (Last 6 weeks)26
    Reflects downloads up to 30 Aug 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Link Recommendation to Augment Influence Diffusion with Provable GuaranteesProceedings of the ACM Web Conference 202410.1145/3589334.3645521(2509-2518)Online publication date: 13-May-2024
    • (2024)Generalized hop‐based approaches for identifying influential nodes in social networksExpert Systems10.1111/exsy.13649Online publication date: 4-Jun-2024
    • (2024)Conformity-aware adoption maximization in competitive social networksNeurocomputing10.1016/j.neucom.2023.127224573(127224)Online publication date: Mar-2024
    • (2024)Time and value aware influence blocking maximization in geo-social networksThe Journal of Supercomputing10.1007/s11227-024-06252-080:14(21149-21178)Online publication date: 6-Jun-2024
    • (2023)Scalable Approximate Butterfly and Bi-triangle Counting for Large Bipartite NetworksProceedings of the ACM on Management of Data10.1145/36267531:4(1-26)Online publication date: 12-Dec-2023
    • (2023)Efficient Algorithm for Budgeted Adaptive Influence Maximization: An Incremental RR-set Update ApproachProceedings of the ACM on Management of Data10.1145/36173281:3(1-26)Online publication date: 13-Nov-2023
    • (2023)A Survey on Influence Maximization: From an ML-Based Combinatorial OptimizationACM Transactions on Knowledge Discovery from Data10.1145/360455917:9(1-50)Online publication date: 18-Jul-2023
    • (2023)Composite Community-Aware Diversified Influence Maximization With Efficient ApproximationIEEE/ACM Transactions on Networking10.1109/TNET.2023.332187032:2(1584-1599)Online publication date: 10-Oct-2023
    • (2023)Influence Maximization within Period under Different Advance Publicity2023 9th International Conference on Applied System Innovation (ICASI)10.1109/ICASI57738.2023.10179532(136-138)Online publication date: 21-Apr-2023
    • (2023)Online conflict resolutionInformation Sciences: an International Journal10.1016/j.ins.2023.119718651:COnline publication date: 1-Dec-2023
    • Show More Cited By

    View Options

    Get Access

    Login options

    Full Access

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    Full Text

    HTML Format

    View this article in HTML Format.

    HTML Format

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media