Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

PLACE Dropout: A Progressive Layer-wise and Channel-wise Dropout for Domain Generalization

Published: 23 October 2023 Publication History
  • Get Citation Alerts
  • Abstract

    Domain generalization (DG) aims to learn a generic model from multiple observed source domains that generalizes well to arbitrary unseen target domains without further training. The major challenge in DG is that the model inevitably faces a severe overfitting issue due to the domain gap between source and target domains. To mitigate this problem, some dropout-based methods have been proposed to resist overfitting by discarding part of the representation of the intermediate layers. However, we observe that most of these methods only conduct the dropout operation in some specific layers, leading to an insufficient regularization effect on the model. We argue that applying dropout at multiple layers can produce stronger regularization effects, which could alleviate the overfitting problem on source domains more adequately than previous layer-specific dropout methods. In this article, we develop a novel layer-wise and channel-wise dropout for DG, which randomly selects one layer and then randomly selects its channels to conduct dropout. Particularly, the proposed method can generate a variety of data variants to better deal with the overfitting issue. We also provide theoretical analysis for our dropout method and prove that it can effectively reduce the generalization error bound. Besides, we leverage the progressive scheme to increase the dropout ratio with the training progress, which can gradually boost the difficulty of training the model to enhance its robustness. Extensive experiments on three standard benchmark datasets have demonstrated that our method outperforms several state-of-the-art DG methods. Our code is available at https://github.com/lingeringlight/PLACEdropout.

    Supplementary Material

    3624015-app (3624015-app.pdf)
    Supplementary material

    References

    [1]
    Yogesh Balaji, Swami Sankaranarayanan, and Rama Chellappa. 2018. Metareg: Towards domain generalization using meta-regularization. In Proceedings of the NeurIPS.
    [2]
    Yoshua Bengio, Jérôme Louradour, Ronan Collobert, and Jason Weston. 2009. Curriculum learning. In Proceedings of the ICML.
    [3]
    Francesco Cappio Borlino, Antonio D’Innocente, and Tatiana Tommasi. 2021. Rethinking domain generalization baselines. In Proceedings of the ICPR.
    [4]
    Fabio M Carlucci, Antonio D’Innocente, Silvia Bucci, Barbara Caputo, and Tatiana Tommasi. 2019. Domain generalization by solving jigsaw puzzles. In Proceedings of the CVPR.
    [5]
    Chaoqi Chen, Jiongcheng Li, Xiaoguang Han, Xiaoqing Liu, and Yizhou Yu. 2022. Compound domain generalization via meta-knowledge encoding. In Proceedings of the CVPR.
    [6]
    Yang Chen, Yu Wang, Yingwei Pan, Ting Yao, Xinmei Tian, and Tao Mei. 2021. A style and semantic memory mechanism for domain generalization. In Proceedings of the ICCV.
    [7]
    Ekin D. Cubuk, Barret Zoph, Jonathon Shlens, and Quoc V. Le. 2020. Randaugment: Practical automated data augmentation with a reduced search space. In Proceedings of the CVPRW.
    [8]
    Qi Dou, Daniel Coelho de Castro, Konstantinos Kamnitsas, and Ben Glocker. 2019. Domain generalization via model-agnostic learning of semantic features. In Proceedings of the NeurIPS.
    [9]
    Antonio D’Innocente and Barbara Caputo. 2018. Domain generalization with domain-specific aggregation modules. In Proceedings of the GCPR.
    [10]
    Xinjie Fan, Qifei Wang, Junjie Ke, Feng Yang, Boqing Gong, and Mingyuan Zhou. 2021. Adversarially adaptive normalization for single domain generalization. In Proceedings of the CVPR.
    [11]
    Golnaz Ghiasi, Tsung-Yi Lin, and Quoc V. Le. 2018. Dropblock: A regularization method for convolutional networks. In Proceedings of the NeurIPS.
    [12]
    Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of the CVPR.
    [13]
    Qibin Hou, Zihang Jiang, Li Yuan, Ming-Ming Cheng, Shuicheng Yan, and Jiashi Feng. 2022. Vision permutator: A permutable MLP-like architecture for visual recognition. TPAMI 45, 1 (2022), 1328–1334.
    [14]
    Saihui Hou and Zilei Wang. 2019. Weighted channel dropout for regularization of deep convolutional neural network. In Proceedings of the AAAI.
    [15]
    Jiaxing Huang, Dayan Guan, Aoran Xiao, and Shijian Lu. 2021. Rda: Robust domain adaptation via fourier adversarial attacking. In Proceedings of the ICCV.
    [16]
    Xun Huang and Serge Belongie. 2017. Arbitrary style transfer in real-time with adaptive instance normalization. In Proceedings of the ICCV.
    [17]
    Zeyi Huang, Haohan Wang, Eric P. Xing, and Dong Huang. 2020. Self-challenging improves cross-domain generalization. In Proceedings of the ECCV.
    [18]
    Juwon Kang, Sohyun Lee, Namyup Kim, and Suha Kwak. 2022. Style neophile: Constantly seeking novel styles for domain generalization. In Proceedings of the CVPR.
    [19]
    Da Li, Yongxin Yang, Yi-Zhe Song, and Timothy M. Hospedales. 2017. Deeper, broader and artier domain generalization. In Proceedings of the ICCV.
    [20]
    Pan Li, Da Li, Wei Li, Shaogang Gong, Yanwei Fu, and Timothy M. Hospedales. 2021. A simple feature augmentation for domain generalization. In Proceedings of the ICCV.
    [21]
    Zekun Li, Lei Qi, Yinghuan Shi, and Yang Gao. 2023. IOMatch: Simplifying open-set semi-supervised learning with joint inliers and outliers utilization. In Proceedings of the ICCV.
    [22]
    Lin Liu, Mingming Zhao, Shanxin Yuan, Wenlong Lyu, Wengang Zhou, Houqiang Li, Yanfeng Wang, and Qi Tian. 2023. Exploring effective mask sampling modeling for neural image compression. Retrieved from https://arXiv:2306.05704
    [23]
    Divyat Mahajan, Shruti Tople, and Amit Sharma. 2021. Domain generalization using causal matching. In Proceedings of the ICML.
    [24]
    Toshihiko Matsuura and Tatsuya Harada. 2020. Domain generalization using a mixture of multiple latent domains. In Proceedings of the AAAI.
    [25]
    Rang Meng, Xianfeng Li, Weijie Chen, Shicai Yang, Jie Song, Xinchao Wang, Lei Zhang, Mingli Song, Di Xie, and Shiliang Pu. 2022. Attention diversification for domain generalization. In Proceedings of the ECCV.
    [26]
    Pietro Morerio, Jacopo Cavazza, Riccardo Volpi, René Vidal, and Vittorio Murino. 2017. Curriculum dropout. In ICCV.
    [27]
    Shruti Nagpal, Maneet Singh, Richa Singh, and Mayank Vatsa. 2020. Attribute aware filter-drop for bias invariant classification. In Proceedings of the CVPRW.
    [28]
    Hyeonseob Nam, HyunJae Lee, Jongchan Park, Wonjun Yoon, and Donggeun Yoo. 2021. Reducing domain gap by reducing style bias. In Proceedings of the CVPR.
    [29]
    Oren Nuriel, Sagie Benaim, and Lior Wolf. 2021. Permuted AdaIN: Reducing the bias towards global statistics in image classification. In Proceedings of the CVPR.
    [30]
    Sinno Jialin Pan and Qiang Yang. 2009. A survey on transfer learning. TKDE 22, 10 (2009), 1345–1359.
    [31]
    Sungheon Park and Nojun Kwak. 2016. Analysis on the dropout effect in convolutional neural networks. In Proceedings of the ACCV.
    [32]
    Jiaxin Qi, Kaihua Tang, Qianru Sun, Xian-Sheng Hua, and Hanwang Zhang. 2022. Class is invariant to context and vice versa: On learning invariance for out-of-distribution generalization. In Proceedings of the ECCV.
    [33]
    Lei Qi, Jiaqi Liu, Lei Wang, Yinghuan Shi, and Xin Geng. 2023. Unsupervised generalizable multi-source person re-identification: A domain-specific adaptive framework. PR 140 (2023).
    [34]
    Yongming Rao, Wenliang Zhao, Zheng Zhu, Jiwen Lu, and Jie Zhou. 2021. Global filter networks for image classification. In Proceedings of the NeurIPS.
    [35]
    Christian Schreckenberger, Christian Bartelt, and Heiner Stuckenschmidt. 2019. iDropout: Leveraging deep taylor decomposition for the robustness of deep neural networks. In Proceedings of the OTM.
    [36]
    Seonguk Seo, Yumin Suh, Dongwan Kim, Geeho Kim, Jongwoo Han, and Bohyung Han. 2020. Learning to optimize domain specific normalization for domain generalization. In Proceedings of the ECCV.
    [37]
    Baifeng Shi, Dinghuai Zhang, Qi Dai, Zhanxing Zhu, Yadong Mu, and Jingdong Wang. 2020. Informative dropout for robust representation learning: A shape-bias perspective. In Proceedings of the ICML.
    [38]
    Qinghongya Shi, Hong-Bo Zhang, Zhe Li, Ji-Xiang Du, Qing Lei, and Jing-Hua Liu. 2022. Shuffle-invariant network for action recognition in videos. TOMM 18, 3 (2022), 1–18.
    [39]
    Yang Shu, Zhangjie Cao, Chenyu Wang, Jianmin Wang, and Mingsheng Long. 2021. Open domain generalization with domain-augmented meta-learning. In Proceedings of the CVPR.
    [40]
    Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. 2014. Dropout: A simple way to prevent neural networks from overfitting. JMLR 15, 1 (2014), 1929–1958.
    [41]
    Jonathan Tompson, Ross Goroshin, Arjun Jain, Yann LeCun, and Christoph Bregler. 2015. Efficient object localization using convolutional networks. In Proceedings of the CVPR.
    [42]
    Antonio Torralba and Alexei A Efros. 2011. Unbiased look at dataset bias. In Proceedings of the CVPR.
    [43]
    Hemanth Venkateswara, Jose Eusebio, Shayok Chakraborty, and Sethuraman Panchanathan. 2017. Deep hashing network for unsupervised domain adaptation. In Proceedings of the CVPR.
    [44]
    Riccardo Volpi, Hongseok Namkoong, Ozan Sener, John Duchi, Vittorio Murino, and Silvio Savarese. 2018. Generalizing to unseen domains via adversarial data augmentation. In Proceedings of the NeurIPS.
    [45]
    Stefan Wager, Sida Wang, and Percy S Liang. 2013. Dropout training as adaptive regularization. In Proceedings of the NeurIPS.
    [46]
    Jindong Wang, Cuiling Lan, Chang Liu, Yidong Ouyang, Tao Qin, Wang Lu, Yiqiang Chen, Wenjun Zeng, and Philip Yu. 2022. Generalizing to unseen domains: A survey on domain generalization. TKDE 35, 8 (2022), 8052–8072.
    [47]
    Shujun Wang, Lequan Yu, Caizi Li, Chi-Wing Fu, and Pheng-Ann Heng. 2020. Learning from extrinsic and intrinsic supervisions for domain generalization. In Proceedings of the ECCV.
    [48]
    Xiran Wang, Jian Zhang, Lei Qi, and Yinghuan Shi. 2023. Generalizable decision boundaries: Dualistic meta-learning for open set domain generalization. In Proceedings of the ICCV.
    [49]
    Yufei Wang, Haoliang Li, Lap-pui Chau, and Alex C Kot. 2021. Embracing the dark knowledge: Domain generalization using regularized knowledge distillation. In Proceedings of the ACM MM.
    [50]
    Yue Wang, Lei Qi, Yinghuan Shi, and Yang Gao. 2022. Feature-based style randomization for domain generalization. TCSVT 32, 8 (2022), 54595–5509.
    [51]
    Zijian Wang, Yadan Luo, Ruihong Qiu, Zi Huang, and Mahsa Baktashmotlagh. 2021. Learning to diversify for single domain generalization. In Proceedings of the ICCV.
    [52]
    Guoqiang Wei, Cuiling Lan, Wenjun Zeng, and Zhibo Chen. 2021. MetaAlign: Coordinating domain alignment and classification for unsupervised domain adaptation. In Proceedings of the CVPR.
    [53]
    Lei Wu, Hefei Ling, Yuxuan Shi, and Baiyan Zhang. 2022. Instance correlation graph for unsupervised domain adaptation. TOMM 18, 1s (2022), 1–23.
    [54]
    Qinwei Xu, Ruipeng Zhang, Ya Zhang, Yanfeng Wang, and Qi Tian. 2021. A fourier-based framework for domain generalization. In Proceedings of the CVPR.
    [55]
    Yifan Xu, Kekai Sheng, Weiming Dong, Baoyuan Wu, Changsheng Xu, and Bao-Gang Hu. 2022. Towards corruption-agnostic robust domain adaptation. TOMM 18, 4 (2022), 1–16.
    [56]
    Zhenzhen Yang, Pengfei Xu, Yongpeng Yang, and Bing-Kun Bao. 2021. A densely connected network based on U-Net for medical image segmentation. TOMM 17, 3 (2021), 1–14.
    [57]
    Xufeng Yao, Yang Bai, Xinyun Zhang, Yuechen Zhang, Qi Sun, Ran Chen, Ruiyu Li, and Bei Yu. 2022. PCL: Proxy-based contrastive learning for domain generalization. In Proceedings of the CVPR.
    [58]
    Matthew D Zeiler and Rob Fergus. 2014. Visualizing and understanding convolutional networks. In Proceedings of the ECCV.
    [59]
    Yuyuan Zeng, Tao Dai, Bin Chen, Shu-Tao Xia, and Jian Lu. 2021. Correlation-based structural dropout for convolutional neural networks. PR 120 (2021).
    [60]
    Dinghuai Zhang, Kartik Ahuja, Yilun Xu, Yisen Wang, and Aaron Courville. 2021. Can subnetwork structure be the key to out-of-distribution generalization? In Proceedings of the ICML.
    [61]
    Jian Zhang, Lei Qi, Yinghuan Shi, and Yang Gao. 2022. More is better: A novel multi-view framework for domain generalization. In Proceedings of the ECCV.
    [62]
    Jian Zhang, Lei Qi, Yinghuan Shi, and Yang Gao. 2022. MVDG: A unified multi-view framework for domain generalization. In Proceedings of the ECCV.
    [63]
    Jian Zhang, Lei Qi, Yinghuan Shi, and Yang Gao. 2023. DomainAdaptor: A novel approach to test-time adaptation. In Proceedings of the ICCV.
    [64]
    Xingxuan Zhang, Peng Cui, Renzhe Xu, Linjun Zhou, Yue He, and Zheyan Shen. 2021. Deep stable learning for out-of-distribution generalization. In Proceedings of the CVPR.
    [65]
    Yabin Zhang, Minghan Li, Ruihuang Li, Kui Jia, and Lei Zhang. 2022. Exact feature distribution matching for arbitrary style transfer and domain generalization. In Proceedings of the CVPR.
    [66]
    Kaiyang Zhou, Ziwei Liu, Yu Qiao, Tao Xiang, and Chen Change Loy. 2021. Domain generalization in vision: A survey. Retrieved from https://arXiv:2103.02503
    [67]
    Kaiyang Zhou, Yongxin Yang, Timothy Hospedales, and Tao Xiang. 2020. Deep domain-adversarial image generation for domain generalisation. In Proceedings of the AAAI.
    [68]
    Kaiyang Zhou, Yongxin Yang, Yu Qiao, and Tao Xiang. 2020. Domain generalization with MixStyle. In Proceedings of the ICLR.

    Cited By

    View all
    • (2024)A Random Focusing Method with Jensen–Shannon Divergence for Improving Deep Neural Network Performance Ensuring Architecture ConsistencyNeural Processing Letters10.1007/s11063-024-11668-z56:4Online publication date: 17-Jun-2024
    • (2024)Reducing Overfitting Risk in Small-Sample Learning with ANN: A Case of Predicting Graduate Admission ProbabilityArtificial Intelligence and Machine Learning10.1007/978-981-97-1277-9_31(404-419)Online publication date: 3-Apr-2024

    Index Terms

    1. PLACE Dropout: A Progressive Layer-wise and Channel-wise Dropout for Domain Generalization

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Transactions on Multimedia Computing, Communications, and Applications
      ACM Transactions on Multimedia Computing, Communications, and Applications  Volume 20, Issue 3
      March 2024
      665 pages
      ISSN:1551-6857
      EISSN:1551-6865
      DOI:10.1145/3613614
      • Editor:
      • Abdulmotaleb El Saddik
      Issue’s Table of Contents

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 23 October 2023
      Online AM: 13 September 2023
      Accepted: 04 September 2023
      Revised: 06 August 2023
      Received: 14 December 2022
      Published in TOMM Volume 20, Issue 3

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. Domain generalization
      2. dropout regularization
      3. overfitting problem
      4. distribution shift

      Qualifiers

      • Research-article

      Funding Sources

      • NSFC
      • Jiangsu Natural Science Foundation Project

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)254
      • Downloads (Last 6 weeks)16
      Reflects downloads up to 27 Jul 2024

      Other Metrics

      Citations

      Cited By

      View all
      • (2024)A Random Focusing Method with Jensen–Shannon Divergence for Improving Deep Neural Network Performance Ensuring Architecture ConsistencyNeural Processing Letters10.1007/s11063-024-11668-z56:4Online publication date: 17-Jun-2024
      • (2024)Reducing Overfitting Risk in Small-Sample Learning with ANN: A Case of Predicting Graduate Admission ProbabilityArtificial Intelligence and Machine Learning10.1007/978-981-97-1277-9_31(404-419)Online publication date: 3-Apr-2024

      View Options

      Get Access

      Login options

      Full Access

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Full Text

      View this article in Full Text.

      Full Text

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media