Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3485447.3511987acmconferencesArticle/Chapter ViewAbstractPublication PagesthewebconfConference Proceedingsconference-collections
research-article

TRACE: A Fast Transformer-based General-Purpose Lossless Compressor

Published: 25 April 2022 Publication History

Abstract

Deep-learning-based compressor has received interests recently due to much improved compression ratio. However, modern approaches suffer from long execution time. To ease this problem, this paper targets on cutting down the execution time of deep-learning-based compressors. Building history-dependencies sequentially (e.g., recurrent neural networks) is responsible for long inference latency. Instead, we introduce transformer into deep learning compressors to build history-dependencies in parallel. However, existing transformer is too heavy in computation and incompatible to compression tasks.
This paper proposes a fast general-purpose lossless compressor, TRACE, by designing a compression-friendly structure based on a single-layer transformer. We first design a new metric to advise the selection part of compression model structures. Byte-grouping and Shared-ffn schemes are further proposed to fully utilize the capacity of the single-layer transformer. These features allow TRACE to achieve competitive compression ratio and a much faster speed. In addition, we further accelerate the compression procedure by designing a controller to reduce the parameter updating overhead. Experiments show that TRACE achieves an overall ∼ 3x speedup while keeps a comparable compression ratio to the state-of-the-art compressors. The source code for TRACE and links to the datasets are available at https://github.com/mynotwo/A-Fast-Transformer-based-General-Purpose-LosslessCompressor.

References

[1]
X. Yu al.2020. Two-Level Data Compression using Machine Learning in Time Series Database. In 2020 IEEE 36th International Conference on Data Engineering (ICDE. 1333–1344.
[2]
Fabrice Bellard. 2019. NNCP: Lossless Data Compression with Neural Networks. (2019). https://bellard.org/nncp/
[3]
Nikolay Bogoychev. 2020. Not all parameters are born equal: Attention is mostly what you need. arXiv preprint arXiv:2010.11859(2020).
[4]
M. Burtscher and P. Ratanaworabhan. 2009. FPC: A High-Speed Compressor for Double-Precision Floating-Point Data. IEEE Trans. Comput. 58, 1 (2009), 18–31.
[5]
[5] Worldwide Quarterly Enterprise Infrastructure Tracker: Buyer and Cloud Deployment.2021. (2021).
[6]
Kyunghyun Cho, Bart Van Merriënboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. 2014. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv preprint arXiv:1406.1078(2014).
[7]
J. Cleary and I. Witten. 1984. Data compression using adaptive coding and partial string matching. IEEE Transacitons on Communications 32, 4 (1984), 396–402.
[8]
John G Cleary and William J Teahan. 1997. Unbounded length contexts for PPM. Comput. J. 40, 2_and_3 (1997), 67–75.
[9]
Y. Collet. 2016. Zstd github repository from facebook.https://github.com/facebook/zstd
[10]
C. Costa, G. Chatzimilioudis, D. Zeinalipour-Yazti, and M. F. Mokbel. 2017. Efficient Exploration of Telco Big Data with Compression and Decaying. In 2017 IEEE 33rd International Conference on Data Engineering (ICDE). 1332–1343.
[11]
C. Costa, G. Chatzimilioudis, D. Zeinalipour-Yazti, and M. F. Mokbel. 2017. SPATE: Compacting and Exploring Telco Big Data. In 2017 IEEE 33rd International Conference on Data Engineering (ICDE). 1419–1420.
[12]
J. Deng, W. Dong, R. Socher, 2009. Imagenet: A large-scale hierarchical image database[C]//2009 IEEE conference on computer vision and pattern recognition. Ieee (2009), 248–255.
[13]
Sebastian Deorowicz. 1985. Silesia Dataset. (1985). http://sun.aei.polsl.pl/sdeor/index.php?page=silesia
[14]
Peter Deutsch. 1996. GZIP file format specification version 4.3. RFC 1952(1996), 1–12. https://doi.org/10.17487/RFC1952
[15]
J. Devlin, M. W. Chang, K. Lee, 2018. Bert: Pre-training of deep bidirectional transformers for language understanding[J]. arXiv. (2018). arxiv:1810.04805preprint.
[16]
DKRZ. 2020. DKRZ. (2020). https://www.research-in-germany.org/en/research-landscape/research-organisations/research-infrastructures/dkrz.html
[17]
Mohit Goyal, Kedar Tatwawadi, Shubham Chandak, and Idoia Ochoa. 2021. DZip: Improved general-purpose loss less compression based on novel neural network modeling. In 2021 Data Compression Conference (DCC). IEEE, 153–162.
[18]
S. Idreos, R. Kaushik, V. Narasayya, and R. Ramamurthy. 2010. Estimating the compression fraction of an index using sampling. In 2010 IEEE 26th International Conference on Data Engineering (ICDE 2010. 441–444.
[19]
David Reinsel John Rydning, John F.Gantz. 2021. 2021–2025: The World Keeps Creating More Data — Now, What Do We Do with It All? (2021). https://www.idc.com/getdoc.jsp?containerId=US46410421
[20]
Friso Kingma, Pieter Abbeel, and Jonathan Ho. 2019. Bit-swap: Recursive bits-back coding for lossless compression with hierarchical latent variables. In International Conference on Machine Learning. PMLR, 3408–3417.
[21]
W Kinsner and RH Greenfield. 1991. The Lempel-Ziv-Welch (LZW) data compression algorithm for packet radio. In [Proceedings] WESCANEX’91. IEEE, 225–229.
[22]
B. Knoll. 2014. CMIX. (2014). http://www.byronknoll.com/cmix.html
[23]
B. Knoll. 2016. Tensorflow-compress. https://github.com/byronknoll/tensorflow-compress
[24]
B. Knoll. 2020. NNCP: Lossless Data Compression with Neural Networks. (2020). https://bellard.org/nncp/
[25]
Byron Knoll and Nando de Freitas. 2012. A machine learning perspective on predictive coding with PAQ8. In 2012 Data Compression Conference. IEEE, 377–386.
[26]
Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu Soricut. 2019. Albert: A lite bert for self-supervised learning of language representations. arXiv preprint arXiv:1909.11942(2019).
[27]
Valerii Likhosherstov, Krzysztof M Choromanski, Jared Quincy Davis, Xingyou Song, and Adrian Weller. 2021. Sub-linear memory: How to make performers slim. Advances in Neural Information Processing Systems 34 (2021).
[28]
Qian Liu, Yiling Xu, and Zhu Li. 2019. DecMac: A Deep Context Model for High Efficiency Arithmetic Coding. In 2019 International Conference on Artificial Intelligence in Information and Communication (ICAIIC). IEEE, 438–443.
[29]
Matt Mahoney. 2006. Large Text Compression Benchmark. (2006). http://mattmahoney.net/dc/text.html
[30]
José L Nunez-Yanez and Vassilios A Chouliaras. 2005. A configurable statistical lossless compression core based on variable order Markov modeling and arithmetic coding. IEEE Trans. Comput. 54, 11 (2005), 1345–1359.
[31]
I. Pavlov. 1999. 7zip: File archiver. (1999). www.7-zip.org
[32]
Karol J. Piczak. 2015. ESC: Dataset for Environmental Sound Classification. (2015). https://doi.org/10.7910/DVN/YDEPUT
[33]
Jack W Rae, Anna Potapenko, Siddhant M Jayakumar, and Timothy P Lillicrap. 2019. Compressive transformers for long-range sequence modelling. arXiv preprint arXiv:1911.05507(2019).
[34]
Claude Elwood Shannon. 1948. A mathematical theory of communication. The Bell system technical journal 27, 3 (1948), 379–423.
[35]
Rajeev Sharma. 2016. Zipline. (2016). https://github.com/opencomputeproject/Project-Zipline
[36]
Y. Tay, M. Dehghani, S. Abnar, 2020. Long Range Arena: A Benchmark for Efficient Transformers. (2020). arxiv:2011.04006preprint.
[37]
James Townsend, Tom Bird, and David Barber. 2019. Practical lossless compression with latent variables using bits back coding. arXiv preprint arXiv:1901.04866(2019).
[38]
Rianne van den Berg, Alexey A Gritsenko, Mostafa Dehghani, Casper Kaae Sønderby, and Tim Salimans. 2020. Idf++: Analyzing and improving integer discrete flows for lossless compression. In International Conference on Learning Representations.
[39]
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. Advances in neural information processing systems 30 (2017).
[40]
Wikipedia. 2020. Amazon Web Serivce. (2020). https://en.wikipedia.org/wiki/Amazon_Web_Services
[41]
Ian H Witten, Radford M Neal, and John G Cleary. 1987. Arithmetic coding for data compression. Commun. ACM 30, 6 (1987), 520–540.
[42]
Jacob Ziv and Abraham Lempel. 1977. A universal algorithm for sequential data compression. IEEE Transactions on information theory 23, 3 (1977), 337–343.

Cited By

View all
  • (2024)A Lossless Compression Technique for the Downlink Control Information Message2024 IEEE 25th International Workshop on Signal Processing Advances in Wireless Communications (SPAWC)10.1109/SPAWC60668.2024.10694087(86-90)Online publication date: 10-Sep-2024
  • (2024)The Likelihood Gain of a Language Model as a Metric for Text Summarization2024 IEEE International Symposium on Information Theory (ISIT)10.1109/ISIT57864.2024.10619426(2044-2049)Online publication date: 7-Jul-2024
  • (2024)ByteZip: Efficient Lossless Compression for Structured Byte Streams Using DNNs2024 International Joint Conference on Neural Networks (IJCNN)10.1109/IJCNN60899.2024.10650523(1-8)Online publication date: 30-Jun-2024
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
WWW '22: Proceedings of the ACM Web Conference 2022
April 2022
3764 pages
ISBN:9781450390965
DOI:10.1145/3485447
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 25 April 2022

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. byte stream
  2. computational efficient model
  3. general-purpose compressor
  4. lossless data compression
  5. neural networks
  6. transformer

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

WWW '22
Sponsor:
WWW '22: The ACM Web Conference 2022
April 25 - 29, 2022
Virtual Event, Lyon, France

Acceptance Rates

Overall Acceptance Rate 1,899 of 8,196 submissions, 23%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)118
  • Downloads (Last 6 weeks)13
Reflects downloads up to 03 Oct 2024

Other Metrics

Citations

Cited By

View all
  • (2024)A Lossless Compression Technique for the Downlink Control Information Message2024 IEEE 25th International Workshop on Signal Processing Advances in Wireless Communications (SPAWC)10.1109/SPAWC60668.2024.10694087(86-90)Online publication date: 10-Sep-2024
  • (2024)The Likelihood Gain of a Language Model as a Metric for Text Summarization2024 IEEE International Symposium on Information Theory (ISIT)10.1109/ISIT57864.2024.10619426(2044-2049)Online publication date: 7-Jul-2024
  • (2024)ByteZip: Efficient Lossless Compression for Structured Byte Streams Using DNNs2024 International Joint Conference on Neural Networks (IJCNN)10.1109/IJCNN60899.2024.10650523(1-8)Online publication date: 30-Jun-2024
  • (2024)TurboLog: A Turbocharged Lossless Compression Method for System Logs via Transformer2024 International Joint Conference on Neural Networks (IJCNN)10.1109/IJCNN60899.2024.10649957(1-10)Online publication date: 30-Jun-2024
  • (2024)GeneFormer: Learned Gene Compression using Transformer-Based Context ModelingICASSP 2024 - 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)10.1109/ICASSP48485.2024.10448360(8035-8039)Online publication date: 14-Apr-2024
  • (2023)Faster and Stronger Lossless Compression with Optimized Autoregressive Framework2023 60th ACM/IEEE Design Automation Conference (DAC)10.1109/DAC56929.2023.10247866(1-6)Online publication date: 9-Jul-2023

View Options

Get Access

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media