Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3485447.3512015acmconferencesArticle/Chapter ViewAbstractPublication PagesthewebconfConference Proceedingsconference-collections
research-article

Massive Text Normalization via an Efficient Randomized Algorithm

Published: 25 April 2022 Publication History

Abstract

Current popular machine learning techniques in natural language processing and data mining rely heavily on high-quality text sources. Nevertheless, real-world text datasets contain a significant amount of spelling errors and improperly punctuated variants where the performance of these models would quickly deteriorate. Moreover, existing text normalization methods are prohibitively expensive to execute over web-scale datasets, can hardly process noisy texts from social networks, or require annotations to learn the corrections in a supervised manner. In this paper, we present Flan (Fast LSH Algorithm for Text Normalization), a scalable randomized algorithm to clean and canonicalize massive text data. Our approach suggests corrections based on the morphology of the words, where lexically similar words are considered the same with high probability. We efficiently handle the pairwise word-to-word comparisons via locality sensitive hashing (LSH). We also propose a novel stabilization process to address the issue of hash collisions between dissimilar words, which is a consequence of the randomized nature of LSH and is exacerbated by the massive scale of real-world datasets. Compared with existing approaches, our method is more efficient, both asymptotically and in empirical evaluations, does not rely on feature engineering, and does not require any annotation. Our experimental results on real-world datasets demonstrate the efficiency and efficacy of Flan. Based on recent advances in densified Minhash, our approach requires much less computational time compared to baseline text normalization techniques on large-scale Twitter and Reddit datasets. In a human evaluation of the quality of the normalization, Flan achieves 5% and 14% improvement against the baselines over the Reddit and Twitter datasets, respectively. Our method also improves performance on Twitter sentiment classification applications and the perturbed GLUE benchmark datasets, where we introduce random errors into the text.

References

[1]
Nir Ailon and Bernard Chazelle. 2009. The Fast Johnson–Lindenstrauss Transform and Approximate Nearest Neighbors. SIAM J. Comput. 39, 1 (2009), 302–322.
[2]
Leena Al-Hussaini. 2017. Experience: Insights into the Benchmarking Data of Hunspell and Aspell Spell Checkers. ACM J. Data Inf. Qual. 8, 3-4 (2017), 13:1–13:10.
[3]
Alexandr Andoni and Piotr Indyk. 2008. Near-optimal hashing algorithms for approximate nearest neighbor in high dimensions. Commun. ACM 51, 1 (2008), 117–122.
[4]
Mark Aronoff. 1994. Morphology by itself: Stems and inflectional classes. Number 22. MIT press.
[5]
AiTi Aw, Min Zhang, Juan Xiao, and Jian Su. 2006. A Phrase-Based Statistical Model for SMS Text Normalization. In ACL. The Association for Computer Linguistics.
[6]
Timothy Baldwin, Marie-Catherine de Marneffe, Bo Han, Young-Bum Kim, Alan Ritter, and Wei Xu. 2015. Shared Tasks of the 2015 Workshop on Noisy User-generated Text. In Proceedings of the Workshop on Noisy User-generated Text. 126–135.
[7]
Béla Bollobás and Paul Erdös. 1976. Cliques in random graphs. In Mathematical Proceedings of the Cambridge Philosophical Society, Vol. 80. Cambridge University Press, 419–427.
[8]
Francesco Bonchi, Ophir Frieder, Franco Maria Nardini, Fabrizio Silvestri, and Hossein Vahabi. 2012. Interactive and context-aware tag spell check and correction. In CIKM. ACM, 1869–1873.
[9]
Andrei Z. Broder and Michael Mitzenmacher. 2001. Completeness and robustness properties of min-wise independent permutations. Random Struct. Algorithms 18, 1 (2001), 18–30.
[10]
Martine Cadot and Joseph di Martino. 2003. A Data Cleaning Solution by Perl Scripts for the KDD Cup 2003 Task 2. SIGKDD Explor. Newsl. 5, 2 (2003), 158–159.
[11]
Jhon Adrián Cerón-Guzmán and Elizabeth León-Guzmán. 2016. Lexical Normalization of Spanish Tweets. In WWW (Companion Volume). ACM, 605–610.
[12]
Beidi Chen, Anshumali Shrivastava, and Rebecca C. Steorts. 2018. Unique entity estimation with application to the Syrian conflict. Ann. Appl. Stat. 12, 2 (06 2018), 1039–1067.
[13]
Monojit Choudhury, Rahul Saraf, Vijit Jain, Animesh Mukherjee, Sudeshna Sarkar, and Anupam Basu. 2007. Investigation and modeling of the structure of texting language. Int. J. Document Anal. Recognit. 10, 3-4 (2007), 157–174.
[14]
Nicole Coddington. 2014. Correction of typographical errors on touch displays. US Patent 8,739,055.
[15]
William W Cohen, Pradeep Ravikumar, Stephen E Fienberg, 2003. A Comparison of String Distance Metrics for Name-Matching Tasks. In IIWeb, Vol. 3. 73–78.
[16]
Richard Cole and Ramesh Hariharan. 2002. Approximate String Matching: A Simpler Faster Algorithm. SIAM J. Comput. 31, 6 (2002), 1761–1782.
[17]
Danish Contractor, Tanveer A. Faruquie, and L. Venkata Subramaniam. 2010. Unsupervised cleansing of noisy text. In COLING (Posters). Chinese Information Processing Society of China, 189–196.
[18]
Graham Cormode and S. Muthukrishnan. 2005. An improved data stream summary: the count-min sketch and its applications. J. Algorithms 55, 1 (2005), 58–75.
[19]
Matthias Damaschk, Tillmann Dönicke, and Florian Lux. 2019. Multiclass Text Classification on Unbalanced, Sparse and Noisy Data. In Proceedings of the First NLPL Workshop on Deep Learning for Natural Language Processing. 58–65.
[20]
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL-HLT. Association for Computational Linguistics, 4171–4186.
[21]
Qin Ding, Hsiang-Fu Yu, and Cho-Jui Hsieh. 2019. A Fast Sampling Algorithm for Maximum Inner Product Search. In AISTATS(Proceedings of Machine Learning Research, Vol. 89). PMLR, 3004–3012.
[22]
Yerai Doval, Jesús Vilares, and Carlos Gómez-Rodríguez. 2020. Towards Robust Word Embeddings for Noisy Texts. Applied Sciences 10, 19 (2020).
[23]
Venkat Gudivada, Amy Apon, and Junhua Ding. 2017. Data quality considerations for big data and machine learning: Going beyond data cleaning and transformations. International Journal on Advances in Software 10, 1 (2017), 1–20.
[24]
Bo Han and Timothy Baldwin. 2011. Lexical Normalisation of Short Text Messages: Makn Sens a #twitter. In ACL. The Association for Computer Linguistics, 368–378.
[25]
Bo Han, Paul Cook, and Timothy Baldwin. 2013. Lexical normalization for social media text. ACM Trans. Intell. Syst. Technol. 4, 1 (2013), 5:1–5:27.
[26]
Aminul Islam and Diana Inkpen. 2009. Real-word spelling correction using Google web 1Tn-gram data set. In CIKM. ACM, 1689–1692.
[27]
Jeff Johnson, Matthijs Douze, and Hervé Jégou. 2021. Billion-Scale Similarity Search with GPUs. IEEE Trans. Big Data 7, 3 (2021), 535–547.
[28]
Nobuhiro Kaji and Masaru Kitsuregawa. 2014. Accurate Word Segmentation and POS Tagging for Japanese Microblogs: Corpus Annotation and Joint Modeling with Lexical Normalization. In EMNLP. ACL, 99–109.
[29]
R. Andrew Kreek and Emilia Apostolova. 2018. Training and Prediction Data Discrepancies: Challenges of Text Classification with Noisy, Historical Data. In NUT@EMNLP. Association for Computational Linguistics, 104–109.
[30]
Jure Leskovec, Anand Rajaraman, and Jeffrey David Ullman. 2014. Mining of Massive Datasets(3nd ed.). Cambridge University Press, USA.
[31]
Ping Li, Art B. Owen, and Cun-Hui Zhang. 2012. One Permutation Hashing. In NIPS. 3122–3130.
[32]
Chen Luo and Anshumali Shrivastava. 2018. Arrays of (locality-sensitive) Count Estimators (ACE): Anomaly Detection on the Edge. In WWW. ACM, 1439–1448.
[33]
Chen Luo and Anshumali Shrivastava. 2018. Arrays of (locality-sensitive) Count Estimators (ACE): Anomaly Detection on the Edge. In WWW. ACM, 1439–1448.
[34]
Dimitrios P. Lyras, Kyriakos N. Sgarbas, and Nikolaos D. Fakotakis. 2007. Using the Levenshtein Edit Distance for Automatic Lemmatization: A Case Study for Modern Greek and English. In ICTAI. IEEE Computer Society, 428–435.
[35]
Valentin Malykh, Varvara Logacheva, and Taras Khakhulin. 2018. Robust Word Vectors: Context-Informed Embeddings for Noisy Texts. In NUT@EMNLP. Association for Computational Linguistics, 54–63.
[36]
Tomás Mikolov, Edouard Grave, Piotr Bojanowski, Christian Puhrsch, and Armand Joulin. 2018. Advances in Pre-Training Distributed Word Representations. In LREC. European Language Resources Association (ELRA).
[37]
Michael Mitzenmacher and Eli Upfal. 2017. Probability and Computing: Randomization and Probabilistic Techniques in Algorithms and Data Analysis(2nd ed.). Cambridge University Press.
[38]
Benjamin Müller, Benoît Sagot, and Djamé Seddah. 2019. Enhancing BERT for Lexical Normalization. In W-NUT@EMNLP. Association for Computational Linguistics, 297–306.
[39]
Priyanka Nigam, Yiwei Song, Vijai Mohan, Vihan Lakshman, Weitian Allen Ding, Ankit Shingavi, Choon Hui Teo, Hao Gu, and Bing Yin. 2019. Semantic Product Search. In KDD. ACM, 2876–2885.
[40]
Peter Norvig. 2009. Natural language corpus data. Beautiful data (2009), 219–242.
[41]
Chris D. Paice. 1994. An Evaluation Method for Stemming Algorithms. In SIGIR. ACM/Springer, 42–50.
[42]
Jeffrey Pennington, Richard Socher, and Christopher D. Manning. 2014. Glove: Global Vectors for Word Representation. In EMNLP. ACL, 1532–1543.
[43]
Tapan Sahni, Chinmay Chandak, Naveen Reddy Chedeti, and Manish Singh. 2017. Efficient Twitter sentiment classification using subjective distant supervision. In COMSNETS. IEEE, 548–553.
[44]
Victor Sanh, Lysandre Debut, Julien Chaumond, and Thomas Wolf. 2019. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. In EMC2@NeurIPS. IEEE, 1–5.
[45]
Kira A. Selby, Yinong Wang, Ruizhe Wang, Peyman Passban, Ahmad Rashid, Mehdi Rezagholizadeh, and Pascal Poupart. 2021. Robust Embeddings Via Distributions. CoRR abs/2104.08420(2021). arxiv:2104.08420
[46]
Victor S. Sheng, Foster J. Provost, and Panagiotis G. Ipeirotis. 2008. Get another label? improving data quality and data mining using multiple, noisy labelers. In KDD. ACM, 614–622.
[47]
Anshumali Shrivastava. 2017. Optimal Densification for Fast and Accurate Minwise Hashing. In ICML(Proceedings of Machine Learning Research, Vol. 70). PMLR, 3154–3163.
[48]
Anshumali Shrivastava and Ping Li. 2014. Densifying One Permutation Hashing via Rotation for Fast Near Neighbor Search. In ICML(JMLR Workshop and Conference Proceedings, Vol. 32). JMLR.org, 557–565.
[49]
Anshumali Shrivastava and Ping Li. 2014. Improved Densification of One Permutation Hashing. In UAI. AUAI Press, 732–741.
[50]
Sujoy Kumar Sikdar, Byungkyu Kang, John O’Donovan, Tobias Hollerer, and Sibel Adal. 2013. Cutting through the noise: Defining ground truth in information credibility on twitter. Human 2, 3 (2013), 151–167.
[51]
Douglas R. Stinson. 1991. Universal Hashing and Authentication Codes. In CRYPTO(Lecture Notes in Computer Science, Vol. 576). Springer, 74–85.
[52]
Dmitry Supranovich and Viachaslau Patsepnia. 2015. IHS_RD: Lexical Normalization for English Tweets. In NUT@IJCNLP. Association for Computational Linguistics, 78–81.
[53]
Rob van der Goot. 2019. MoNoise: A Multi-lingual and Easy-to-use Lexical Normalization Tool. In ACL. Association for Computational Linguistics, 201–206.
[54]
Rob van der Goot, Alan Ramponi, Tommaso Caselli, Michele Cafagna, and Lorenzo De Mattei. 2020. Norm It! Lexical Normalization for Italian and Its Downstream Effects for Dependency Parsing. In LREC. European Language Resources Association, 6272–6278.
[55]
Michael Völske, Martin Potthast, Shahbaz Syed, and Benno Stein. 2017. TL;DR: Mining Reddit to Learn Automatic Summarization. In NFiS@EMNLP. Association for Computational Linguistics, 59–63.
[56]
Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel R. Bowman. 2019. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding. In ICLR (Poster). OpenReview.net.
[57]
Kang Zhao, Hongtao Lu, and Jincheng Mei. 2014. Locality Preserving Hashing. In AAAI. AAAI Press, 2874–2881.

Cited By

View all
  • (2024)Language Identification and Normalization Techniques for Code-Mixed Text2024 Sixth International Conference on Computational Intelligence and Communication Technologies (CCICT)10.1109/CCICT62777.2024.00077(435-441)Online publication date: 19-Apr-2024

Index Terms

  1. Massive Text Normalization via an Efficient Randomized Algorithm
        Index terms have been assigned to the content through auto-classification.

        Recommendations

        Comments

        Information & Contributors

        Information

        Published In

        cover image ACM Conferences
        WWW '22: Proceedings of the ACM Web Conference 2022
        April 2022
        3764 pages
        ISBN:9781450390965
        DOI:10.1145/3485447
        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

        Sponsors

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        Published: 25 April 2022

        Permissions

        Request permissions for this article.

        Check for updates

        Author Tags

        1. Lexical Normalization
        2. Locality-Sensitive Hashing
        3. Natural Language Processing

        Qualifiers

        • Research-article
        • Research
        • Refereed limited

        Conference

        WWW '22
        Sponsor:
        WWW '22: The ACM Web Conference 2022
        April 25 - 29, 2022
        Virtual Event, Lyon, France

        Acceptance Rates

        Overall Acceptance Rate 1,899 of 8,196 submissions, 23%

        Contributors

        Other Metrics

        Bibliometrics & Citations

        Bibliometrics

        Article Metrics

        • Downloads (Last 12 months)31
        • Downloads (Last 6 weeks)2
        Reflects downloads up to 30 Aug 2024

        Other Metrics

        Citations

        Cited By

        View all
        • (2024)Language Identification and Normalization Techniques for Code-Mixed Text2024 Sixth International Conference on Computational Intelligence and Communication Technologies (CCICT)10.1109/CCICT62777.2024.00077(435-441)Online publication date: 19-Apr-2024

        View Options

        Get Access

        Login options

        View options

        PDF

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader

        HTML Format

        View this article in HTML Format.

        HTML Format

        Media

        Figures

        Other

        Tables

        Share

        Share

        Share this Publication link

        Share on social media