Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3592979.3593401acmconferencesArticle/Chapter ViewAbstractPublication PagespascConference Proceedingsconference-collections
research-article
Public Access

Scaling Resolution of Gigapixel Whole Slide Images Using Spatial Decomposition on Convolutional Neural Networks

Published: 26 June 2023 Publication History

Abstract

Gigapixel images are prevalent in scientific domains ranging from remote sensing, and satellite imagery to microscopy, etc. However, training a deep learning model at the natural resolution of those images has been a challenge in terms of both, overcoming the resource limit (e.g. HBM memory constraints), as well as scaling up to a large number of GPUs. In this paper, we trained Residual neural Networks (ResNet) on 22,528 x 22,528-pixel size images using a distributed spatial decomposition method on 2,304 GPUs on the Summit Supercomputer. We applied our method on a Whole Slide Imaging (WSI) dataset from The Cancer Genome Atlas (TCGA) database. WSI images can be in the size of 100,000 x 100,000 pixels or even larger, and in this work we studied the effect of image resolution on a classification task, while achieving state-of-the-art AUC scores. Moreover, our approach doesn't need pixel-level labels, since we're avoiding patching from the WSI images completely, while adding the capability of training arbitrary large-size images. This is achieved through a distributed spatial decomposition method, by leveraging the non-block fat-tree interconnect network of the Summit architecture, which enabled GPU-to-GPU direct communication. Finally, detailed performance analysis results are shown, as well as a comparison with a data-parallel approach when possible.

References

[1]
Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248--255. Ieee, 2009.
[2]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. CoRR, abs/1512.03385, 2015.
[3]
Peter Jin, Boris Ginsburg, and Kurt Keutzer. Spatially parallel convolutions, 2018.
[4]
Amir Gholami, Ariful Azad, Peter Jin, Kurt Keutzer, and Aydin Buluc. Integrated model, batch and domain parallelism in training neural networks. 2017.
[5]
Nikoli Dryden, Naoya Maruyama, Tom Benson, Tim Moon, Marc Snir, and Brian Van Essen. Improving strong-scaling of cnn training by exploiting finer-grained parallelism, 2019.
[6]
Le Hou, Youlong Cheng, Noam Shazeer, Niki Parmar, Yeqing Li, Panagiotis Korfiatis, Travis M. Drucker, Daniel J. Blezek, and Xiaodan Song. High resolution medical image analysis with spatial partitioning, 2019.
[7]
Sudip K. Seal, Seung-Hwan Lim, Dali Wang, Jacob Hinkle, Dalton Lunga, and Aristeidis Tsaris. Toward large-scale image segmentation on summit. In 49th International Conference on Parallel Processing - ICPP, ICPP '20, New York, NY, USA, 2020. Association for Computing Machinery.
[8]
Aristeidis Tsaris, Jacob Hinkle, Dalton Lunga, and Philipe Ambrozio Dias. Distributed training for high resolution images: A domain and spatial decomposition approach. In 2021 IEEE/ACM Redefining Scalability for Diversely Heterogeneous Architectures Workshop (RSDHA), pages 27--33, 2021.
[9]
Chi Long Chen, Chi Long Chen, Chi-Chung Chen, Wei-Hsiang Yu, Szu Hua Chen, Yu Chan Chang, Tai I. Hsu, Michael Hsiao, Chao-Yuan Yeh, and Cheng Yu Chen. An annotation-free whole-slide training approach to pathological classification of lung cancer types using deep learning. Nature Communications, 12(1):1193--1193, 2021.
[10]
Aoxiao Zhong and Quanzheng Li. Team hms-mgh-ccds method1. 2018.
[11]
Oded Maron and Tomás Lozano-Pérez. A framework for multiple-instance learning. Advances in neural information processing systems, 10, 1997.
[12]
Ming Y Lu, Drew FK Williamson, Tiffany Y Chen, Richard J Chen, Matteo Barbieri, and Faisal Mahmood. Data-efficient and weakly supervised computational pathology on whole-slide images. Nature Biomedical Engineering, 5(6):555--570, 2021.
[13]
Kausik Das, Sailesh Conjeti, Abhijit Guha Roy, Jyotirmoy Chatterjee, and Debdoot Sheet. Multiple instance learning of deep convolutional neural networks for breast histopathology whole slide classification. In 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), pages 578--581. IEEE, 2018.
[14]
Jiawen Yao, Xinliang Zhu, and Junzhou Huang. Deep multi-instance learning for survival prediction from whole slide images. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pages 496--504. Springer, 2019.
[15]
Abtin Riasatian. Kimianet: Training a deep network for histopathology using high-cellularity. Master's thesis, University of Waterloo, 2020.
[16]
Richard J Chen, Chengkuan Chen, Yicong Li, Tiffany Y Chen, Andrew D Trister, Rahul G Krishnan, and Faisal Mahmood. Scaling vision transformers to gigapixel images via hierarchical self-supervised learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 16144--16155, 2022.
[17]
Jingwei Zhang, Xin Zhang, Ke Ma, Rajarsi Gupta, Joel Saltz, Maria Vakalopoulou, and Dimitris Samaras. Gigapixel whole-slide images classification using locally supervised learning. In Lecture Notes in Computer Science, pages 192--201. Springer Nature Switzerland, 2022.
[18]
Ming Y. Lu, Drew F. K. Williamson, Tiffany Y. Chen, Richard J. Chen, Matteo Barbieri, and Faisal Mahmood. Data efficient and weakly supervised computational pathology on whole slide images, 2020.
[19]
Zhuchen Shao, Hao Bian, Yang Chen, Yifeng Wang, Jian Zhang, Xiangyang Ji, and Yongbing Zhang. Transmil: Transformer based correlated multiple instance learning for whole slide image classification, 2021.
[20]
Richard J. Chen, Chengkuan Chen, Yicong Li, Tiffany Y. Chen, Andrew D. Trister, Rahul G. Krishnan, and Faisal Mahmood. Scaling vision transformers to gigapixel images via hierarchical self-supervised learning, 2022.
[21]
H. Pinckaers, B. van Ginneken, and G. Litjens. Streaming convolutional neural networks for end-to-end learning with multi-megapixel images. IEEE Transactions on Pattern Analysis amp; Machine Intelligence, 44(03):1581--1590, mar 2022.
[22]
Verónica G. Vergara Larrea, Wayne Joubert, Michael J. Brim, Reuben D. Budiardja, Don Maxwell, Matt Ezell, Christopher Zimmer, Swen Boehm, Wael Elwasif, Sarp Oral, Chris Fuson, Daniel Pelfrey, Oscar Hernandez, Dustin Leverman, Jesse Hanley, Mark Berrill, and Arnold Tharrington. Scaling the summit: Deploying the world's fastest supercomputer. In Michèle Weiland, Guido Juckeland, Sadaf Alam, and Heike Jagode, editors, High Performance Computing, pages 330--351, Cham, 2019. Springer International Publishing.
[23]
Nitish Shirish Keskar, Dheevatsa Mudigere, Jorge Nocedal, Mikhail Smelyanskiy, and Ping Tak Peter Tang. On large-batch training for deep learning: Generalization gap and sharp minima. CoRR, abs/1609.04836, 2016.
[24]
Andrew Brock, Soham De, Samuel L. Smith, and Karen Simonyan. High-performance large-scale image recognition without normalization, 2021.

Cited By

View all
  • (2024)Infer-HiRes: Accelerating Inference for High-Resolution Images with Quantization and Distributed Deep LearningPractice and Experience in Advanced Research Computing 2024: Human Powered Computing10.1145/3626203.3670548(1-9)Online publication date: 17-Jul-2024

Index Terms

  1. Scaling Resolution of Gigapixel Whole Slide Images Using Spatial Decomposition on Convolutional Neural Networks

        Recommendations

        Comments

        Information & Contributors

        Information

        Published In

        cover image ACM Conferences
        PASC '23: Proceedings of the Platform for Advanced Scientific Computing Conference
        June 2023
        274 pages
        ISBN:9798400701900
        DOI:10.1145/3592979
        Publication rights licensed to ACM. ACM acknowledges that this contribution was authored or co-authored by an employee, contractor or affiliate of the United States government. As such, the Government retains a nonexclusive, royalty-free right to publish or reproduce this article, or to allow others to do so, for Government purposes only.

        Sponsors

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        Published: 26 June 2023

        Permissions

        Request permissions for this article.

        Check for updates

        Author Tags

        1. distributed deep learning
        2. model parallelism
        3. spatial decomposition
        4. medical imaging
        5. convolutional neural networks

        Qualifiers

        • Research-article

        Funding Sources

        Conference

        PASC '23
        Sponsor:

        Acceptance Rates

        Overall Acceptance Rate 109 of 221 submissions, 49%

        Contributors

        Other Metrics

        Bibliometrics & Citations

        Bibliometrics

        Article Metrics

        • Downloads (Last 12 months)89
        • Downloads (Last 6 weeks)17
        Reflects downloads up to 31 Jan 2025

        Other Metrics

        Citations

        Cited By

        View all
        • (2024)Infer-HiRes: Accelerating Inference for High-Resolution Images with Quantization and Distributed Deep LearningPractice and Experience in Advanced Research Computing 2024: Human Powered Computing10.1145/3626203.3670548(1-9)Online publication date: 17-Jul-2024

        View Options

        View options

        PDF

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader

        Login options

        Figures

        Tables

        Media

        Share

        Share

        Share this Publication link

        Share on social media