Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.5555/3432601.3432629dlproceedingsArticle/Chapter ViewAbstractPublication PagescasconConference Proceedingsconference-collections
research-article

Deep learning approaches to classify the relevance and sentiment of news articles to the economy

Published: 10 November 2020 Publication History

Abstract

We consider a text classification task over an open source dataset involving news snippets and their relevance to the US economy. Text classification and sentiment analysis have been performed using nine different classifiers among which three are the traditional machine learning models, namely, support vector machine, extreme gradient boosting and logistic regression, and six neural network-based methods. The neural net frameworks include long short-term memory (LSTM), bidirectional long short-term memory (BiLSTM) and an ensemble of one dimensional convolution network (1D CNN) with LSTM/BiLSTM. Both word-to-vector and term-frequency inverse-document-frequency vectors are used in our analysis with text and sentiment classification tasks. A detailed comparative study is provided to assess the relative performance of different classification approaches. It is observed that the ensemble with 1D CNN performs better in both binary and multiclass classifications. Specifically, in the multinomial sentiment classification, 1D CNN with BiLSTM has the best performance as opposed to 1D CNN with LSTM in the binary text classification. BiLSTM architecture which incorporates the backward dependencies turns out as superior to LSTM by a margin of 30% in multiclass classification even though the considered dataset is small and inherently challenging. Further analysis to evaluate the impact of successive increases in percentage of augmented data reveals that such augmentation has a limit up to 180% in this dataset beyond which the performance starts decreasing.

References

[1]
[n.d.]. Introduction to news analytics. https://www.eventstudytools.com/introduction-news-analytics. Accessed: 2020-06-22.
[2]
[n.d.]. News feeds, analytics, and indices. https://www.refinitiv.com/en/products/world-news-data. Accessed: 2020-06-22.
[3]
Martín Abadi, Ashish Agarwal, Paul Barham, Eugene Brevdo, and et al. 2015. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. http://tensorflow.org/Software available from tensorflow.org.
[4]
Xuemei Bai. 2018. Text classification based on LSTM and attention. In 2018 Thirteenth International Conference on Digital Information Management (ICDIM). IEEE, 29--32.
[5]
Steven Bird, Ewan Klein, and Edward Loper. 2009. Natural language processing with Python: analyzing text with the natural language toolkit. "O'Reilly Media, Inc.".
[6]
Piotr Bojanowski, Edouard Grave, Armand Joulin, and Tomas Mikolov. 2016. Enriching word vectors with subword information. arXiv (2016), preprint arXiv:1607.04606.
[7]
Tiago Carneiro, Raul Victor Medeiros Da Nóbrega, Thiago Nepomuceno, GuiBin Bian, Victor Hugo C De Albuquerque, and Pedro Pedrosa Reboucas Filho. 2018. Performance analysis of google colaboratory as a tool for accelerating deep learning applications. IEEE Access 6 (2018), 61677--61685.
[8]
Nitesh V Chawla, Kevin W Bowyer, Lawrence O Hall, and W Philip Kegelmeyer. 2002. Smote: synthetic minority over-sampling technique. Journal of artificial intelligence research 16 (2002), 321--357.
[9]
Tianqi Chen and Carlos Guestrin. 2016. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining. 785--794.
[10]
François Chollet et al. 2015. Keras. https://keras.io.
[11]
Poonam Choudhari and S Veena Dhari. 2017. Sentiment Analysis and Machine Learning Based Sentiment Classification: A Review. International Journal of Advanced Research in Computer Science 8, 3 (2017).
[12]
Edward Collins, Nikolai Rozanov, and Bingbing Zhang. 2018. Evolutionary Data Measures: Understanding the Difficulty of Text Classification Tasks. arXiv preprint arXiv:1811.01910 (2018).
[13]
Alexis Conneau, Holger Schwenk, Loic Barrault, and Yann Lecun. 2017. Very deep convolutional networks for text classification. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics 1 (2017), 1107--1116.
[14]
Corinna Cortes and Vladimir Vapnik. 1995. Support-vector networks. Machine learning 20, 3 (1995), 273--297.
[15]
CrowdFlower. 2015. Economic News Article Tone, https://data.world/crowdflower/economic-news-article-tone. (Dec 2015). Dec 2015.
[16]
Sanjiv R Das. 2011. News analytics: Framework, techniques and metrics. The Handbook of News Analytics in Finance 2 (2011).
[17]
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv:cs.CL/1810.04805
[18]
Tadayoshi Fushiki. 2011. Estimation of prediction error by using K-fold cross-validation. Statistics and Computing 21, 2 (2011), 137--146.
[19]
Maya R Gupta, SamyBengio, and Jason Weston. 2014. Training highly multiclass classifiers. The Journal of Machine Learning Research 15, 1 (2014), 1461--1492.
[20]
Geoffrey E Hinton and Russ R Salakhutdinov. 2009. Replicated softmax: an undirected topic model. In Advances in neural information processing systems. 1607--1614.
[21]
Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long short-term memory. Neural computation 9, 8 (1997), 1735--1780.
[22]
Matthew Honnibal and Ines Montani. 2017. spaCy 2: Natural language understanding with Bloom embeddings, convolutional neural networks and incremental parsing. (2017). To appear.
[23]
K Sparck Jones. 2001. Natural language processing: a historical review. University of Cambridge (2001), 2--10.
[24]
Armand Joulin, Edouard Grave, Piotr Bojanowski, and Tomas Mikolov. 2016. Bag of Tricks for Efficient Text Classification. arXiv preprint arXiv:1607.01759 (2016).
[25]
Tarek Kanan, Odai Sadaqa, Amal Aldajeh, Hanadi Alshwabka, Shadi AlZu'bi, Mohammed Elbes, Bilal Hawashin, Mohammad A Alia, et al. 2019. A review of natural language processing and machine learning tools used to analyze arabic social media. In 2019 IEEE Jordan International Joint Conference on Electrical Engineering and Information Technology (JEEIT). IEEE, 622--628.
[26]
Ayman E Khedr, Nagwa Yaseen, et al. 2017. Predicting stock market behavior using data mining technique and news sentiment analysis. International Journal of Intelligent Systems and Applications 9, 7 (2017), 22.
[27]
Max R Kimbrough, Steven O Kimbrough, and Priscilla Murphy. 2011. On using text analytics for event studies. In Proceedings of the 13th International Conference on Artificial Intelligence and Law. 209--218.
[28]
David G Kleinbaum, K Dietz, M Gail, Mitchel Klein, and Mitchell Klein. 2002. Logistic regression. Springer.
[29]
Yann LeCun, Yoshua Bengio, and Geoffrey Hinton. 2015. Deep learning. Nature 521, 7553 (2015), 436--444.
[30]
Yann LeCun, Bernhard Boser, John S Denker, Donnie Henderson, Richard E Howard, Wayne Hubbard, and Lawrence D Jackel. 1989. Backpropagation applied to handwritten zip code recognition. Neural Computation 1, 4 (1989), 541--551.
[31]
Gang Liu and Jiabao Guo. 2019. Bidirectional LSTM with attention mechanism and convolutional layer for text classification. Neurocomputing 337 (2019), 325--338.
[32]
Yifei Lu, Yanghui Rao, Jun Yang, and Jian Yin. 2018. Incorporating Lexicons into LSTM for sentiment classification. In 2018 International joint conference on neural networks (IJCNN). IEEE, 1--7.
[33]
G Harry Mc Laughlin. 1969. Smog grading-a new readability formula. Journal of reading 12 (1969), 639--646.
[34]
Mary L McHugh. 2012. Interrater reliability: the kappa statistic. Biochemia medica: Biochemia medica 22, 3 (2012), 276--282.
[35]
R Monika, S Deivalakshmi, and B Janet. 2019. Sentiment Analysis of US Airlines Tweets Using LSTM/RNN. In 2019 IEEE 9th International Conference on Advanced Computing (IACC). IEEE, 92--95.
[36]
Travis E Oliphant. 2006. A guide to NumPy. Vol. 1. Trelgol Publishing USA.
[37]
Daniel W Otter, Julian R Medina, and Jugal K Kalita. 2020. A Survey of the Usages of Deep Learning for Natural Language Processing. IEEE Transactions on Neural Networks and Learning Systems (2020).
[38]
The pandas development team. 2020. pandas-dev/pandas: Pandas.
[39]
Fabian Pedregosa, Gaël Varoquaux, Alexandre Gramfort, Vincent Michel, Bertrand Thirion, Olivier Grisel, Mathieu Blondel, Peter Prettenhofer, Ron Weiss, Vincent Dubourg, et al. 2011. Scikit-learn: Machine learning in Python. the Journal of machine Learning research 12 (2011), 2825--2830.
[40]
F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. 2011. Scikit-learn: Machine Learning in Python. Journal of Machine Learning Research 12 (2011), 2825--2830.
[41]
Jeffrey Pennington, Richard Socher, and Christopher D. Manning. 2014. GloVe: Global Vectors for Word Representation. In Empirical Methods in Natural Language Processing (EMNLP). 1532--1543. http://www.aclweb.org/anthology/D14-1162
[42]
David Martin Powers. 2011. Evaluation: from precision, recall and f-measure to roc, informedness, markedness and correlation. Journal of Machine Learning Technologies 2, 1 (2011), 37--63.
[43]
Nicolas Pröllochs, Stefan Feuerriegel, and Dirk Neumann. 2015. Enhancing sentiment analysis of financial news by detecting negation scopes. In 2015 48th Hawaii International Conference on System Sciences. IEEE, 959--968.
[44]
Juan Ramos et al. 2003. Using tf-idf to determine word relevance in document queries. In Proceedings of the first instructional conference on machine learning, Vol. 242. Piscataway, NJ, 133--142.
[45]
Gabriele Ranco, Darko Aleksovski, Guido Caldarelli, Miha Grčar, and Igor Mozetič. 2015. The effects of Twitter sentiment on stock price returns. PloS one 10, 9 (2015).
[46]
Gabriele Ranco, Ilaria Bordino, Giacomo Bormetti, Guido Caldarelli, Fabrizio Lillo, and Michele Treccani. 2016. Coupling news sentiment with web browsing data improves prediction of intra-day price dynamics. PLoS one 11, 1 (2016).
[47]
Jürgen Schmidhuber. 2015. Deep learning in neural networks: An overview. Neural networks 61 (2015), 85--117.
[48]
Mike Schuster and Kuldip K Paliwal. 1997. Bidirectional recurrent neural networks. IEEE transactions on Signal Processing 45, 11 (1997), 2673--2681.
[49]
Achira Jeewaka Shamal, Rankothge Gishan Hiranya Pemathilake, Sachith Paramie Karunathilake, and Gamage Upeksha Ganegoda. 2018. Sentiment Analysis using Token2Vec and LSTMs: User Review Analyzing Module. In 2018 18th International Conference on Advances in ICT for Emerging Regions (ICTer). IEEE, 48--53.
[50]
Claude Elwood Shannon. 2001. A mathematical theory of communication. ACM SIGMOBILE Mobile Computing and Communications Review 5 (2001), 3--55.
[51]
Richard Socher, Eric H Huang, Jeffrey Pennin, Christopher D Manning, and Andrew Y Ng. 2011. Dynamic pooling and unfolding recursive autoencoders for paraphrase detection. In Advances in neural information processing systems. 801--809.
[52]
Sahar Sohangir, Dingding Wang, Anna Pomeranets, and Taghi M Khoshgoftaar. 2018. Big Data: Deep Learning for financial sentiment analysis. Journal of Big Data 5, 1 (2018), 3.
[53]
Peter D Turney. 2002. Thumbs up or thumbs down?: semantic orientation applied to unsupervised classification of reviews. In Proceedings of the 40th annual meeting on association for computational linguistics. Association for Computational Linguistics, 417--424.
[54]
Shuo Xu. 2018. Bayesian Naïve Bayes classifiers to text classification. Journal of Information Science 44, 1 (2018), 48--59.
[55]
Dani Yogatama, Chris Dyer, Wang Ling, and Phil Blunsom. 2017. Generative and discriminative text classification with recurrent neural networks. arXiv (2017), preprint arXiv:1703.01898.
[56]
Tom Young, Devamanyu Hazarika, Soujanya Poria, and Erik Cambria. 2018. Recent trends in deep learning based natural language processing. IEEE Computational intelligenCe magazine 13, 3 (2018), 55--75.
[57]
Xiang Zhang, Junbo Zhao, and Yann LeCun. 2015. Character-level convolutional networks for text classification. Advances in neural information processing systems (2015), pages 649--657.

Recommendations

Comments

Information & Contributors

Information

Published In

cover image DL Hosted proceedings
CASCON '20: Proceedings of the 30th Annual International Conference on Computer Science and Software Engineering
November 2020
297 pages

Sponsors

  • IBM Centre for Advanced Studies (CAS)
  • IBM Canada: IBM Canada

Publisher

IBM Corp.

United States

Publication History

Published: 10 November 2020

Author Tags

  1. 1D CNN
  2. tf-idf
  3. BiLSTM
  4. LSTM
  5. NLP
  6. US economy
  7. news snippet
  8. sentiment analysis
  9. text classification

Qualifiers

  • Research-article

Acceptance Rates

Overall Acceptance Rate 24 of 90 submissions, 27%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 65
    Total Downloads
  • Downloads (Last 12 months)19
  • Downloads (Last 6 weeks)1
Reflects downloads up to 25 Jan 2025

Other Metrics

Citations

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media