Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

Neural Topic Modeling via Discrete Variational Inference

Published: 16 February 2023 Publication History

Abstract

Topic models extract commonly occurring latent topics from textual data. Statistical models such as Latent Dirichlet Allocation do not produce dense topic embeddings readily integratable into neural architectures, whereas earlier neural topic models are yet to fully take advantage of the discrete nature of the topic space. To bridge this gap, we propose a novel neural topic model, Discrete-Variational-Inference-based Topic Model (DVITM), which learns dense topic embeddings homomorphic to word embeddings via discrete variational inference. The model also views words as mixtures of topics and digests embedded input text. Quantitative and qualitative evaluations empirically demonstrate the superior performance of DVITM compared to important baseline models. In the end, case studies on text generation from a discrete space and aspect-aware item recommendation are presented to further illustrate the power of our model in downstream tasks.
Appendices

A Baseline Topic Modeling Architectures

Fig. 8.
Fig. 8. ProdLDA with gaussian distribution re-parameterization trick.
Fig. 9.
Fig. 9. Embedded topic model with Gaussian distribution re-parameterization trick (ETM).
Fig. 10.
Fig. 10. ProdLDA with Gumbel-Softmax re-parameterization trick (ProdLDA\(^\prime\)).
Fig. 11.
Fig. 11. ETM with Gumbel-Softmax re-parameterization trick (ETM\(^\prime\)).

B Latent Topics Identified by Various Models

Table 9.
ModelTopic IDTopic CoherenceTop 10 nearest words
ProdLDA00.4658[minnesota, angeles, toronto, san, montreal, vancouver, ottawa, stanley, calgary, louis]
10.3945[finish, play, team, hitter, offense, smith, defensive, ice, tie, hit]
20.5129[christian, god, scripture, interpretation, jesus, resurrection, teaching, doctrine, existence, holy]
30.3626[troops, israel, border, turks, army, israeli, fire, arab, civilian, minority]
40.5116[ram, windows, quadra, fine, speed, scsi, faster, microsoft, apple, external]
50.3740[entry, char, compile, db, file, section, variable, contest, distribution, remark]
60.5557[christian, scripture, god, teaching, resurrection, existence, jesus, christianity, doctrine, biblical]
70.3890[ibm, interface, transfer, external, virtual, cpu, ram, dec, path, default]
80.4294[fire, car, safety, btw, andy, surrender, country, rider, cold, stupid]
90.4889[wiretap, escrow, drug, agency, warrant, clipper, illegal, encryption, crime, country]
100.4657[wiretap, encryption, escrow, nsa, chip, clipper, pat, cheaper, scheme, car]
110.4874[turks, armenian, army, turkish, russian, armenia, united, mountain, armenians, director]
120.4587[music, fine, crash, ram, sl, external, simm, honda, wm, hd]
130.4824[bmw, car, bike, rider, baseball, btw, cop, ball, hit, motorcycle]
140.4628[satellite, mission, distribute, nasa, spacecraft, km, space, earth, distribution, module]
150.5309[fine, ram, windows, apple, amp, simm, external, scsi, crash, quadra]
160.4720[agency, encryption, cryptography, telephone, wiretap, enforcement, des, privacy, distribution, encrypt]
170.3659[neighbor, christian, heart, harm, building, scripture, jesus, daughter, woman, holy]
180.4507[scsi, ram, hd, external, mhz, meg, scsus, ide, fine, bus]
190.5172[sl, wm, ram, mi, external, hd, connector, mg, mb, mw]
ProdLDA\(^\prime\)00.3261[first, series, information, next, team, get, san, mailing, use, also]
10.5680[buf, hp, exit, oname, char, printf, toolkit, mov, bh, saturn]
20.6156[wings, flyers, puck, pit, leafs, scripture, que, hitter, resurrection, penalty]
30.5557[lebanese, armenians, apartment, troops, azerbaijan, hitter, wings, coach, armenian, armenia]
40.3214[key, strong, secret, warrant, people, la, phone, des, algorithm, rights]
50.3737[power, little, league, game, head, go, team, make, turn, sport]
60.4611[lebanese, muslim, batf, bike, lebanon, arabs, troops, witness, apartment, massacre]
70.5425[coach, braves, leafs, hitter, rangers, tor, playoff, stanley, puck, pat]
80.5674[scsus, sl, mb, quadra, hd, meg, pd, byte, motherboard, workstation]
90.6024[satan, ford, simm, resurrection, bmw, bike, doctrine, gear, quadra, scripture]
100.6974[encrypt, encryption, wiretap, escrow, crypto, cipher, rsa, anonymous, pgp, cryptography]
110.3734[vehicle, police, chip, technique, transmission, dealer, cop, car, traffic, radar]
120.6243[mi, eus, spacecraft, mw, moon, mg, orbit, ah, rg, ax]
130.3331[want, buy, like, drive, solution, control, people, know, driver, anyone]
140.4621[armenia, morality, proceed, revelation, bmw, verse, turks, soul, resurrection, bike]
150.5764[xlib, toolkit, pixel, xterm, visual, motif, turbo, microsoft, printf, meg]
160.4564[bhj, spacecraft, bh, eus, wm, byte, device, ripem, wire, digital]
170.4783[mhz, mb, adapter, bhj, scsi, scsus, windows, wm, hd, isa]
180.3429[escape, window, conflict, see, document, terrorist, let, shell, motif, cambridge]
190.4948[militia, arab, palestinian, homicide, lebanese, arabs, turks, armenia, troops, armenians]
ETM00.3489[get, know, one, say, think, like, see, thing, people, time]
10.3254[use, may, make, case, many, also, part, however, system, president]
20.4505[god, jesus, christian, say, believe, bible, one, christ, make, belief]
30.3682[gun, people, child, kill, drug, crime, weapon, police, case, claim]
40.5255[datum, space, db, launch, output, hus, widget, dod, nasa, sun]
50.3390[file, use, program, send, available, list, code, email, please, line]
60.4139[hockey, team, new, division, san, canada, nhl, toronto, york, gm]
70.3439[new, look, buy, price, good, sell, include, package, offer, like]
80.3851[car, power, use, drive, speed, engine, wire, water, fast, low]
90.4651[game, year, play, win, team, player, season, go, good, run]
100.3312[information, make, get, please, mail, use, go, file, help, take]
110.3378[book, first, one, time, science, study, earth, author, history, find]
120.3516[university, group, internet, information, computer, fax, center, year, call, research]
130.3628[thanks, david, john, appreciate, steve, mark, wonder, jim, mike, michael]
140.3749[write, article, post, question, read, opinion, ask, please, yes, answer]
150.3397[period, la, pt, vs, de, van, pp, cal, power, second]
160.3607[key, government, law, use, encryption, state, chip, public, right, security]
170.3354[go, take, back, one, day, put, get, right, also, call]
180.3987[use, drive, system, window, card, run, windows, disk, problem, image]
190.4854[israel, people, war, israeli, jews, turkish, armenians, country, armenian, government]
\(\text{ETM}^{\prime }\)00.3749[use, wiring, connector, code, line, voltage, get, ground, find, might]
10.3876[game, read, news, times, know, beat, hear, braves, go, back]
20.5399[apple, macintosh, amiga, graphics, processor, pc, modem, computer, server, printer]
30.3503[know, want, get, really, say, see, ca, never, think, tell]
40.3341[think, know, name, really, something, people, many, feel, thing, like]
50.4001[gm, please, want, get, maybe, know, somebody, make, lot, hey]
60.3833[book, verse, copy, write, author, manual, guide, reader, story, edition]
70.4091[tax, billion, budget, pay, fee, federal, package, dollar, money, please]
80.3644[get, know, want, think, tell, go, really, take, find, ask]
90.3416[god, know, say, think, see, like, make, use, want, one]
100.3316[like, make, use, write, go, many, get, help, need, know]
110.3827[league, nhl, game, team, list, mail, international, use, first, new]
120.4393[peace, palestinian, visit, state, foreign, israel, islamic, arab, conference, muslim]
130.5090[fax, computer, pc, electronic, systems, hardware, software, graphics, nt, server]
140.3841[go, let, oh, hey, please, eat, red, hang, stay, waco]
150.4863[belief, christianity, religion, islam, faith, people, muslims, sense, believe, god]
160.3508[shell, institute, justice, research, fund, professor, microsoft, minority, secretary, panel]
170.3367[god, see, use, know, say, go, want, like, one, need]
180.4485[federal, constitution, law, amendment, state, act, government, enforcement, authority, right]
190.5186[christianity, religion, armenians, islam, jews, religious, turks, turkish, muslims, turkey]
Table 9. 20 Latent Topics Identified by Baseline Models from 20Ng

C Baseline Models in Case Studies

C.1 VAE-LM: Baseline Model in Case Study 1

VAE-LM [4] (Figure 12) leverages the Gaussian RT from VAE in the language modeling task. Specifically, the model combines following three components:
Fig. 12.
Fig. 12. Variational autoencoder language model.
(1)
LSTM [16] encoder network, Enc(.),
(2)
Gaussian RT from VAE (see Section 2.2), and
(3)
LSTM decoder network, Dec(.).
Given a example text of the form \({\bf x} = (x_1, x_2 \ldots x_n)\) where each \(x_i\) comes from a fixed set of vocab, an encoder (component 1) is utilized to encode the text (see Figure 12 in Appendix A):
\begin{equation} {\bf h} = \text{Enc}({\bf x}), \end{equation}
(31)
where h represents the encoded representation of text x.
At this point, similarly to ProdLDA (see Section 2.3), several feed-forward linear networks (FF1 and FF2) are utilized to produce a K-dimension mean vector \(\boldsymbol {\mu }\) and a log-variance vector \(\boldsymbol {\sigma }^2\) of the same dimension. Then, the RT (component 2) [20] is applied to sample a standard Gaussian vector of dimension \(K\). At this point, the sampled latent representation, \({\bf z}\), is formulated by using Equation (2).
Last, a LSTM-decoder (component 3) is applied to reconstruct the original text conditioned on the latent representation, \({\bf z}\).
\begin{equation} {\bf x}^{\prime \prime } = \text{Dec}({\bf z}), \end{equation}
(32)
where \({\bf x}^{\prime \prime }\) represents the generated text.
Consequently, one can utilize the loss function in Equation (1) to train the neural-network parameters. For VAE-LM, various implementations of encoder and decoder exist. In this case study, we consider the implementation where encoder (Enc(.)) digests both word embeddings and character embeddings available in the input text, x. Additionally, the decoder (Dec(.)) ingests latent representation \({\bf z}\) at every step of generation as shown in Figure 12.

C.2 AARM: Baseline Model in Case Study 2

AARM [11] (Figure 13) considers the following three modules to compute a rating, \(\hat{y}_{uv}\), for a user u and a product v. Let us assume a pre-trained aspect embedding matrix, \({\bf W}_A\) \(\in\)\(\mathbb {R}\) \(^{d_a x A}\), a learnable user embedding matrix, \({\bf W}_U\) \(\in\) \(\mathbb {R}\) \(^{d_g x U}\), and a learnable product embedding matrix, \({\bf W}_V\) \(\in\) \(\mathbb {R}\) \(^{d_g x V}\), where A, U, and V represents total number of aspects, users, and products, respectively, and \(d_a\), \(d_g\) correspond to the dimension of any aspect, user, and product representation. Now, for a user u and for a product v, let us further assume that \(M_u\) and \(M_v\) aspects are extracted (by Sentires), which results in projections \({\bf F}_u\) \(\in\) \(\mathbb {R}\) \(^{d_a x M_u}\) and \({\bf F}_v\) \(\in\) \(\mathbb {R}\) \(^{d_a x M_v}\). Next, the following modules are implemented to predict the rating (\(\hat{y}_{uv}\)) between user u and item v.
Fig. 13.
Fig. 13. AARM.
(1)
Aspect Interaction
Aspect Embedding Transformation
\begin{equation} \begin{split} {\bf C}_u = \frac{{\bf W}_{trans}{\bf F}_u}{||{\bf W}_{trans}{\bf F}_u||}, {\bf C}_v = \frac{{\bf W}_{trans}{\bf F}_v}{||{\bf W}_{trans}{\bf F}_v||}. \end{split} \end{equation}
(33)
Aspect Interaction Layer
\begin{equation} {\bf F}_{AI}(u, v) = \lbrace {\bf C}_u \odot {\bf C}_v({\bf X}_u{\bf X}_v)\rbrace . \end{equation}
(34)
Aspect-Level Attentive Pooling Layer
\begin{equation} \begin{split} \hat{\beta }_{u,v} &= {\bf w}_{att_{1}}^{T} ({\bf C}_u \odot {\bf C}_v({\bf X}_u{\bf X}_v)) \\ \beta _{u,v} &= \frac{\exp (\hat{\beta }_{u,v})}{\sum _{v}\exp (\hat{\beta }_{u,v})}\\ {\bf H}_{u} &= \sum _{v}\beta _{u,v}({\bf C}_u \odot {\bf C}_v({\bf X}_u{\bf X}_v)). \end{split} \end{equation}
(35)
User-Level Attentive Pooling Layer
\begin{equation} \begin{split} {\bf X}_{v, u} &= {\bf g}_{v} \odot {\bf C}_{u} \\ {\bf g}_{v} &= \sum _{v} {\bf C}_v \\ \hat{\gamma }_{u, v} &= {\bf w}_{att_{2}}^{T} {\bf X}_{v, u} \\ \gamma _{u,v} &= \frac{\exp (\hat{\gamma }_{u,v})}{\sum _{u}\exp (\hat{\gamma }_{u,v})}. \end{split} \end{equation}
(36)
Output
\begin{equation} {\bf y}_{A} (u, v) = \sum _{u} \gamma _{u,v}{\bf H}_{u}, \end{equation}
(37)
where
\({\bf W}_{trans}\) \(\in\) \(\mathbb {R}\)\(^{d_a x d_a}\), \({\bf w}_{att_{1}}\) \(\in\) \(\mathbb {R}\) \(^{d_a}\), \({\bf w}_{att_{2}}\) \(\in\) \(\mathbb {R}\) \(^{d_a}\) are learnable parameters.
\({\bf X}_u\) \(\in\) {0, 1} and \({\bf X}_u\) \(\in\) {0, 1} are masking indicators that are used to mask <PAD> tokens.
\({\bf y}_{A} (u, v)\) \(\in\) \(\mathbb {R}\) \(^{d_a}\) is the output of aspect interaction module.
\(\odot\) represents a element-wise multiplication operator.
(2)
Global Interaction
\begin{equation} {\bf y}_{G}(u, v) = {\bf p}_{u} \odot {\bf q}_{v}, \end{equation}
(38)
where
\({\bf p}_{u}\) corresponds to the representation of user u from embedding matrix \({\bf W}_U\).
\({\bf q}_{v}\) corresponds to the representation of product v from embedding matrix \({\bf W}_V\).
\({\bf y}_{G} (u, v)\) \(\in\) \(\mathbb {R}\) \(^{d_g}\) is the output of aspect interaction module.
(3)
Output Layer
\begin{equation} \hat{y}(u, v) = {\bf W}_{out} \texttt {concat}({\bf y}_{G}(u, v), {\bf y}_{A}(u, v)), \end{equation}
(39)
where
\({\bf W}_{out}\) \(\in\) \(\mathbb {R}\) \(^{1 x (d_a + d_g)}\) is a learnable parameter.
\(\hat{y} (u, v)\) represents the overall satisfaction of user u toward product v.
For simplicity, we summarize Equations (33)–(39) as below:
\begin{equation} \hat{y}(u, v) = \text{AARM}({\bf p}_{u}, {\bf q}_{v}, {\bf F}_{u}, {\bf F}_{v}). \end{equation}
(40)
Loss Function: According to Reference [11], AARM utilizes the Bayesian Personalized Ranking method where given a user u, a triple is formulated (u, \(v^{+}\), \(v^{-}\)) to perform the training. \(v^{+}\) refers to the product that user u has purchased whereas \(v^{-}\) represents an unpurchased item. Additionally, the positive pair (u, \(v^{+}\)), in conjunction with a negative pair (u, \(v^{-}\)), is drawn from the rating set R. In summary, AARM utilizes the following loss function to train the whole model,
\begin{equation} \begin{split} L_{bpr} &#x0026;= \frac{-1}{|R|} \sum _{(u, v^{+}) \in R} \log (\sigma (\hat{y}(u, v^{+})- \hat{y}(u, v^{-}))), \end{split} \end{equation}
(41)
\begin{equation} \begin{split} L_{aarm} &#x0026;= L_{bpr} + \lambda * \left(\frac{||{\bf W}_{U}^{2}||}{|{\bf W}_{U}|} + \frac{||{\bf W}_{V}^{2}||}{|{\bf W}_{V}|} + \frac{||{\bf W}_{out}^{2}||}{|{\bf W}_{out}|}\right), \end{split} \end{equation}
(42)
where second term in total loss \(L_{aarm}\) represents the regularization term to avoid overfitting.

References

[1]
Bengio Yoshua, Courville Aaron, and Vincent Pascal. 2013. Representation learning: A review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 35, 8 (2013), 1798–1828.
[2]
David M. Blei and Jon D. Mcauliffe. 2008. Supervised topic models. In Advances in Neural Information Processing Systems Volume 20. 121–128.
[3]
David M. Blei, Andrew Y. Ng, and Michael I. Jordan. 2003. Latent dirichlet allocation. J. Mach. Learn. Res. 3, Jan (2003), 993–1022.
[4]
Samuel R. Bowman, Luke Vilnis, Oriol Vinyals, Andrew M. Dai, Rafal Jozefowicz, and Samy Bengio. 2016. Generating sentences from a continuous space. In Proceedings of the 20th SIGNLL Conference on Computational Natural Language Learning. 10–21.
[5]
Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam Mccandlish, Alec Radford, Ilya Sutskever, and Dario Amodei Openai. 2020. Language Models are Few-Shot Learners. Technical Report.
[6]
Scott Deerwester, Susan T. Dumais, George W. Furnas, Thomas K. Landauer, and Richard Harshman. 1988. Improving information retrieval with latent semantic indexing. In Proceedings of the Annual Meeting of the American Society for Information Science 25. 36–40.
[7]
Adji B. Dieng, Francisco J. R. Ruiz, and David M. Blei. 2020. Topic modeling in embedding spaces. Trans. Assoc. Comput. Ling. 8 (2020), 439–453.
[8]
Katayoun Farrahi and Daniel Gatica-Perez. 2011. Discovering routines from large-scale human locations using probabilistic topic models. ACM Trans. Intell. Syst. Technol. 2, 1 (2011), 1–27.
[9]
Yang Gao, Yuefeng Li, Raymond Y K Lau, Yue Xu, Abul Bashar, and Md Abul Bashar. 2017. Finding semantically valid and relevant topics by association-based topic selection model. ACM Trans. Intell. Syst. Technol. 9, 3 (2017).
[10]
Thomas L. Griffiths and Mark Steyvers. 2004. Finding scientific topics. Proc. Natl. Acad. Sci. U.S.A. 101 (2004), 5228–5235.
[11]
Xinyu Guan, Zhiyong Cheng, Xiangnan He, Yongfeng Zhang, Zhibo Zhu, Qinke Peng, and Tat-Seng Chua. 2019. Attentive aspect modeling for review-aware recommendation. ACM Trans. Inf. Syst. 37, 3 (2019), 1–27.
[12]
Amulya Gupta and Zhu Zhang. 2021. Vector-quantization-based topic modeling. ACM Trans. Intell. Syst. Technol. 12, 3 (2021), 1–30.
[13]
Junxian He, Zhiting Hu, Berg Taylor Kirkpatrick, Ying Huang, and P. Eric Xing. 2017. Efficient correlated topic modeling with topic embedding. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 225–233.
[14]
Karl Moritz Hermann, Tomas Kocisky, Edward Grefenstette, Lasse Espeholt, Will Kay, Mustafa Suleyman, and Phil Blunsom. 2015. Teaching machines to read and comprehend. In Advances in Neural Information Processing Systems. 1693–1701.
[15]
I. Higgins, L. Matthey, A. Pal, C. Burgess, and X. Glorot. 2017. beta-vae: Learning basic visual concepts with a constrained variational framework. In Proceedings of the International Conference on Learning Representations.
[16]
Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long short-term memory. Neural Comput. 9 (1997), 1735–1780.
[17]
Eric Jang, Google Brain, Shixiang Gu, and Ben Poole. 2017. Categorical reparameterization with gumbel-softmax. In Proceedings of the International Conference on Learning Representations.
[18]
Shuning Jin, Sam Wiseman, Karl Stratos, and Karen Livescu. 2020. Discrete latent variable representations for low-resource text classification. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 4831–4842.
[19]
Lang Ken. 1995. NewsWeeder: Learning to filter netnews. In Proceedings of the 12th International Conference on International Conference on Machine Learning. 331–339.
[20]
Diederik P. Kingma and Max Welling. 2014. Auto-encoding variational bayes. stat 1050 (2014), 1.
[21]
Hugo Larochelle and Stanislas Lauly. 2012. A neural autoregressive topic model. In Advances in Neural Information Processing Systems. 2708–2716.
[22]
Shaohua Li, Tat-Seng Chua, Jun Zhu, and Chunyan Miao. 2016. Generative topic embedding: A continuous representation of documents. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics. 666–675.
[23]
Shaohua Li, Jun Zhu, and Chunyan Miao. 2017. PSDVec: A toolbox for incremental and scalable word embedding. Neurocomputing 237 (2017), 405–409.
[24]
Yang Liu, Zhiyuan Liu, Tat-Seng Chua, and Maosong Sun. 2015. Topical word embeddings. In Proceedings of the 29th AAAI Conference on Artificial Intelligence. 2418–2424.
[25]
André F. T. Martins, Andre Martins, and Ramón F. Astudillo. 2016. From softmax to sparsemax: A sparse model of attention and multi-label classification. In Proceedings of the International Conference on Machine Learning. 1614–1623.
[26]
Arya D. McCarthy, Xian Li, Jiatao Gu, and Ning Dong. 2020. Addressing posterior collapse with mutual information for improved variational neural machine translation. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 8512–8525.
[27]
Yishu Miao, Edward Grefenstette, and Phil Blunsom. 2017. Discovering discrete latent topics with neural variational inference. In Proceedings of the International Conference on Machine Learning. 2410–2419.
[28]
David Newman, Jey Han Lau, Karl Grieser, and Timothy Baldwin. 2010. Automatic evaluation of topic coherence. In Human Language Technologies: The Annual Conference of the North American Chapter of the Association for Computational Linguistics. 100–108.
[29]
Alberto Paccanaro and E. Geoffrey Hinton. 2001. Learning distributed representations of concepts using linear relational embedding. IEEE Trans. Knowl. Data Eng. 13 (2001), 232–244.
[30]
Jeffrey Pennington, Richard Socher, and Christopher Manning. 2014. Glove: Global vectors for word representation. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP’14). Association for Computational Linguistics, 1532–1543.
[31]
Nhathai Phan, Javid Ebrahimi, David Kil, Brigitte Piniewski, and Dejing Dou. 2016. Topic-aware physical activity propagation with temporal dynamics in a health social network. ACM Trans. Intell. Syst. Technol. 8, 2 (2016), 1–20.
[32]
Mehdi Rezaee and Francis Ferraro. 2020. A discrete variational recurrent topic model without the reparametrization trick. In Advances in Neural Information Processing Systems. 13831–13843.
[33]
Michael Röder, Andreas Both, and Alexander Hinneburg. 2015. Exploring the space of topic coherence measures. In Proceedings of the International Conference on Web Search and Data Mining.
[34]
Ruslan Salakhutdinov and Geoffrey Hinton. 2009. Replicated softmax: An undirected topic model. In Advances in Neural Information Processing Systems 22. 1607–1614.
[35]
Akash Srivastava and Charles Sutton. 2017. Autoencoding variational inference for topic models. In Proceedings of the International Conference on Learning Representations.
[36]
Yee Whye Teh, Michael I. Jordan, Matthew J. Beal, and David M. Blei. 2006. Hierarchical dirichlet processes. J. Am. Stat. Assoc. 101 (2006), 1566–1581.
[37]
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Kaiser Lukasz, and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems, I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (Eds.). Curran Associates, Inc., 5998–6008.
[38]
Zhijun Yin, Liangliang Cao, Quanquan Gu, and Jiawei Han. 2012. Latent community topic analysis: Integration of community discovery with topic modeling. ACM Trans. Intell. Syst. Technol. 3, 4 (92012).

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Transactions on Intelligent Systems and Technology
ACM Transactions on Intelligent Systems and Technology  Volume 14, Issue 2
April 2023
430 pages
ISSN:2157-6904
EISSN:2157-6912
DOI:10.1145/3582879
  • Editor:
  • Huan Liu
Issue’s Table of Contents

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 16 February 2023
Online AM: 25 November 2022
Accepted: 12 October 2022
Revised: 26 July 2022
Received: 14 December 2021
Published in TIST Volume 14, Issue 2

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Topic modeling
  2. neural models
  3. discrete variational inference
  4. recommendation

Qualifiers

  • Research-article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 388
    Total Downloads
  • Downloads (Last 12 months)133
  • Downloads (Last 6 weeks)6
Reflects downloads up to 10 Nov 2024

Other Metrics

Citations

View Options

Get Access

Login options

Full Access

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Full Text

View this article in Full Text.

Full Text

HTML Format

View this article in HTML Format.

HTML Format

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media