Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

Frequency-aware Camouflaged Object Detection

Published: 23 March 2023 Publication History

Abstract

Camouflaged object detection (COD) is important as it has various potential applications. Unlike salient object detection (SOD), which tries to identify visually salient objects, COD tries to detect objects that are visually very similar to the surrounding background. We observe that recent COD methods try to fuse features from different levels using some context aggregation strategies originally developed for SOD. Such an approach, however, may not be appropriate for COD as these existing context aggregation strategies are good at detecting distinctive objects while weakening the features from less discriminative objects. To address this problem, we propose in this article to exploit frequency learning to suppress the confusing high-frequency texture information, to help separate camouflaged objects from their surrounding background, and a frequency-based method, called FBNet, for camouflaged object detection. Specifically, we design a frequency-aware context aggregation (FACA) module to suppress high-frequency information and aggregate multi-scale features from a frequency perspective, an adaptive frequency attention (AFA) module to enhance the features of the learned important frequency components, and a gradient-weighted loss function to guide the proposed method to pay more attention to contour details. Experimental results show that our model outperforms relevant state-of-the-art methods.

References

[1]
Hassan Abdulameer, Jakub Błaszczyk, Tomasz Depta, Adam Kornacki, and Przemysław Kozieł.2018. Animal Camouflage Analysis: Chameleon Database. Unpublished Manuscript 2, 6 (2018), 7.
[2]
Kai Chen, Jiangmiao Pang, Jiaqi Wang, Yu Xiong, Xiaoxiao Li, Shuyang Sun, Wansen Feng, Ziwei Liu, Jianping Shi, Wanli Ouyang, et al. 2019. Hybrid task cascade for instance segmentation. In Proceedings of the CVPR.
[3]
Liang-Chieh Chen, George Papandreou, Iasonas Kokkinos, Kevin Murphy, and Alan L. Yuille. 2017. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Transactions on Pattern Analysis and Machine Intelligence 40, 4 (2017), 834–848.
[4]
Yao Chen, Xu Qianqian, Cong Runmin, and Huang. Qingming. 2020. Global context-aware progressive aggregation network for salient object detection. In Proceedings of the AAAI.
[5]
Zuyao Chen, Qianqian Xu, Runmin Cong, and Qingming Huang. 2020. Global context-aware progressive aggregation network for salient object detection. In Proceedings of the AAAI.
[6]
Henghui Ding, Xudong Jiang, Bing Shuai, Ai Qun Liu, and Gang Wang. 2018. Context contrasted feature and gated multi-scale aggregation for scene segmentation. In Proceedings of the CVPR.
[7]
Max Ehrlich and Larry S. Davis. 2019. Deep residual learning in the jpeg transform domain. In Proceedings of the ICCV. 3484–3493.
[8]
Deng-Ping Fan, Ming-Ming Cheng, Yun Liu, Tao Li, and Ali Borji. 2017. Structure-measure: A new way to evaluate foreground maps. In Proceedings of the ICCV.
[9]
Deng-Ping Fan, Cheng Gong, Yang Cao, Bo Ren, Ming-Ming Cheng, and Ali Borji. 2018. Enhanced-alignment measure for binary foreground map evaluation. In Proceedings of the IJCAI.
[10]
Deng-Ping Fan, Ge-Peng Ji, Guolei Sun, Ming-Ming Cheng, Jianbing Shen, and Ling Shao. 2020. Camouflaged object detection. In Proceedings of the CVPR.
[11]
Zhengyang Feng, Qianyu Zhou, Qiqi Gu, Xin Tan, Guangliang Cheng, Xuequan Lu, Jianping Shi, and Lizhuang Ma. 2022. DMT: Dynamic mutual training for semi-supervised learning. Pattern Recognition (2022), 108777. DOI:
[12]
Shiming Ge, Xin Jin, Qiting Ye, Zhao Luo, and Qiang Li. 2018. Image editing by object-aware optimal boundary searching and mixed-domain composition. Computational Visual Media 4, 1 (2018), 71–82.
[13]
Robert Geirhos, Patricia Rubisch, Claudio Michaelis, Matthias Bethge, Felix A. Wichmann, and Wieland Brendel. 2019. ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. In Proceedings of the ICLR. Retrieved from https://openreview.net/forum?id=Bygh9j09KX.
[14]
Huankang Guan, Jiaying Lin, and Rynson W. H. Lau. 2022. Learning semantic associations for mirror detection. In Proceedings of the CVPR.
[15]
Kaiming He, Georgia Gkioxari, Piotr Dollár, and Ross Girshick. 2017. Mask r-cnn. In Proceedings of the ICCV.
[16]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2015. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the ICCV. 1026–1034.
[17]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of the CVPR.
[18]
Jie Hu, Li Shen, and Gang Sun. 2018. Squeeze-and-excitation networks. In Proceedings of the CVPR.
[19]
Zhaojin Huang, Lichao Huang, Yongchao Gong, Chang Huang, and Xinggang Wang. 2019. Mask scoring r-cnn. In Proceedings of the CVPR.
[20]
Trung-Nghia Le, Tam V. Nguyen, Zhongliang Nie, Minh-Triet Tran, and Akihiro Sugimoto. 2019. Anabranch network for camouflaged object segmentation. CVIU 184 (2019), 45–56.
[21]
Aixuan Li, Jing Zhang, Yunqiu Lv, Bowen Liu, Tong Zhang, and Yuchao Dai. 2021. Uncertainty-aware joint salient object and camouflaged object detection. In Proceedings of the CVPR.
[22]
Yu Li and Michael S. Brown. 2014. Single image layer separation using relative smoothness. In Proceedings of the CVPR.
[23]
Jiaying Lin, Zebang He, and Rynson W. H. Lau. 2021. Rich context aggregation with reflection prior for glass surface detection. In Proceedings of the CVPR.
[24]
Jiaying Lin, Guodong Wang, and Rynson W. H. Lau. 2020. Progressive mirror detection. In Proceedings of the CVPR.
[25]
Tsung-Yi Lin, Piotr Dollár, Ross Girshick, Kaiming He, Bharath Hariharan, and Serge Belongie. 2017. Feature pyramid networks for object detection. In Proceedings of the CVPR.
[26]
Jiang-Jiang Liu, Qibin Hou, Ming-Ming Cheng, Jiashi Feng, and Jianmin Jiang. 2019. A simple pooling-based design for real-time salient object detection. In Proceedings of the CVPR.
[27]
Nian Liu, Junwei Han, and Ming-Hsuan Yang. 2018. Picanet: Learning pixel-wise contextual attention for saliency detection. In Proceedings of the CVPR.
[28]
Songtao Liu, Di Huang, et al. 2018. Receptive field block net for accurate and fast object detection. In Proceedings of the ECCV. 385–400.
[29]
Wei Liu, Andrew Rabinovich, and Alexander C Berg. 2015. Parsenet: Looking wider to see better. arXiv:1506.04579. Retrieved from https://arxiv.org/abs/1506.04579.
[30]
Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining Guo. 2021. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the ICCV (2021).
[31]
Wenjie Luo, Yujia Li, Raquel Urtasun, and Richard Zemel. 2016. Understanding the effective receptive field in deep convolutional neural networks. In Proceedings of the NeurIPS.
[32]
Yunqiu Lyu, Jing Zhang, Yuchao Dai, Aixuan Li, Bowen Liu, Nick Barnes, and Deng-Ping Fan. 2021. Simultaneously localize, segment and rank the camouflaged objects. In Proceedings of the CVPR.
[33]
Ran Margolin, Lihi Zelnik-Manor, and Ayellet Tal. 2014. How to evaluate foreground maps? In Proceedings of the CVPR.
[34]
Haiyang Mei, Ge-Peng Ji, Ziqi Wei, Xin Yang, Xiaopeng Wei, and Deng-Ping Fan. 2021. Camouflaged object segmentation with distraction mining. In Proceedings of the CVPR.
[35]
Haiyang Mei, Xin Yang, Yang Wang, Yuanyuan Liu, Shengfeng He, Qiang Zhang, Xiaopeng Wei, and Rynson W. H. Lau. 2020. Don’t hit me! glass detection in real-world scenes. In Proceedings of the CVPR.
[36]
Yuxin Pan, Yiwang Chen, Qiang Fu, Ping Zhang, and Xin Xu. 2011. Study on the camouflaged target detection method based on 3D convexity. Modern Applied Science 5, 4 (2011), 152.
[37]
Youwei Pang, Xiaoqi Zhao, Lihe Zhang, and Huchuan Lu. 2020. Multi-scale interactive network for salient object detection. In Proceedings of the CVPR.
[38]
Xuebin Qin, Zichen Zhang, Chenyang Huang, Chao Gao, Masood Dehghan, and Martin Jagersand. 2019. BASNet: Boundary-aware salient object detection. In Proceedings of the CVPR.
[39]
Zequn Qin, Pengyi Zhang, Fei Wu, and Xi Li. 2020. FcaNet: Frequency channel attention networks. arXiv:2012.11879. Retrieved from https://arxiv.org/abs/2012.11879.
[40]
Jinming Su, Jia Li, Yu Zhang, Changqun Xia, and Yonghong. Tian. 2019. Selectivity or invariance: Boundary-aware salient object detection. In Proceedings of the ICCV.
[41]
Xin Tan, Jiaying Lin, Ke Xu, Chen Pan, Lizhuang Ma, and Rynson W. H. Lau. 2022. Mirror detection with the visual chirality cue. IEEE Transactions on Pattern Analysis and Machine Intelligence (2022). DOI:
[42]
Xin Tan, Ke Xu, Ying Cao, Yiheng Zhang, Lizhuang Ma, and Rynson W. H. Lau. 2021. Night-time scene parsing with a large real dataset. IEEE Transactions on Image Processing 30 (2021), 9085–9098.
[43]
Tiantian Wang, Ali Borji, Lihe Zhang, Pingping Zhang, and Huchuan Lu. 2017. A stagewise refinement model for detecting salient objects in images. In Proceedings of the ICCV.
[44]
Jun Wei, Shuhui Wang, and Qingming Huang. 2020. F3Net: Fusion, feedback and focus for salient object detection. In Proceedings of the AAAI.
[45]
Zhe Wu, Li Su, and Qingming Huang. 2019. Cascaded partial decoder for fast and accurate salient object detection. In Proceedings of the CVPR.
[46]
Zhe Wu, Li Su, and Qingming Huang. 2019. Stacked cross refinement network for edge-aware salient object detection. In Proceedings of the ICCV.
[47]
Yingyue Xu, Dan Xu, Xiaopeng Hong, Wanli Ouyang, Rongrong Ji, Min Xu, and Guoying Zhao. 2019. Structured modeling of joint deep feature and prediction refinement for salient object detection. In Proceedings of the ICCV.
[48]
Feng Xue, Chengxi Yong, Shan Xu, Hao Dong, Yuetong Luo, and Wei Jia. 2016. Camouflage performance analysis and evaluation framework based on features fusion. Multimedia Tools and Applications 75, 7 (2016), 4065–4082.
[49]
Chuan Yang, Lihe Zhang, Huchuan Lu, Xiang Ruan, and Ming-Hsuan Yang. 2013. Saliency detection via graph-based manifold ranking. In Proceedings of the CVPR.
[50]
Shuquan Ye, Dongdong Chen, Songfang Han, and Jing Liao. 2021. Learning with noisy labels for robust point cloud segmentation. In Proceedings of the ICCV (2021).
[51]
Jianqin Yin, Yanbin Han, Wendi Hou, and Jinping Li. 2011. Detection of the mobile object with camouflage color under dynamic background based on optical flow. Procedia Engineering 15 (2011), 2201–2205. DOI:
[52]
Qing Zhang, Gelin Yin, Yongwei Nie, and Wei-Shi Zheng. 2020. Deep camouflage images. In Proceedings of the AAAI.
[53]
Xiaoning Zhang, Tiantian Wang, Jinqing Qi, Huchuan Lu, and Gang Wang. 2018. Progressive attention guided recurrent network for salient object detection. In Proceedings of the CVPR.
[54]
Hengshuang Zhao, Jianping Shi, Xiaojuan Qi, Xiaogang Wang, and Jiaya Jia. 2017. Pyramid scene parsing network. In Proceedings of the CVPR.
[55]
Jia-Xing Zhao, Jiang-Jiang Liu, Deng-Ping Fan, Yang Cao, Jufeng Yang, and Ming-Ming Cheng. 2019. EGNet: Edge guidance network for salient object detection. In Proceedings of the ICCV.
[56]
Ting Zhao and Xiangqian Wu. 2019. Pyramid feature attention network for saliency detection. In Proceedings of the ICCV. 3085–3094.
[57]
Ting Zhao and Xiangqian Wu. 2019. Pyramid feature attention network for saliency detection. In Proceedings of the CVPR.
[58]
Xiaoqi Zhao, Youwei Pang, Lihe Zhang, Huchuan Lu, and Lei Zhang. 2020. Suppress and balance: A simple gated network for salient object detection. In Proceedings of the ECCV.
[59]
Zongwei Zhou, Md Mahfuzur Rahman Siddiquee, Nima Tajbakhsh, and Jianming Liang. 2019. UNet++: Redesigning skip connections to exploit multiscale features in image segmentation. IEEE Transactions on Medical Imaging 39, 6 (2019), 1856–1867.
[60]
Wangjiang Zhu, Shuang Liang, Yichen Wei, and Jian Sun. 2014. Saliency optimization from robust background detection. In Proceedings of the CVPR.
[61]
Xueyan Zou, Fanyi Xiao, Zhiding Yu, and Yong Jae Lee. 2020. Delving deeper into anti-aliasing in ConvNets. In Proceedings of the BMVC.

Cited By

View all
  • (2025)Frequency-Guided Spatial Adaptation for Camouflaged Object DetectionIEEE Transactions on Multimedia10.1109/TMM.2024.352168127(72-83)Online publication date: 2025
  • (2025)CoNet: A Consistency-Oriented Network for Camouflaged Object SegmentationIEEE Transactions on Circuits and Systems for Video Technology10.1109/TCSVT.2024.346246535:1(287-299)Online publication date: Jan-2025
  • (2024)A new benchmark for camouflaged object detection: RGB-D camouflaged object detection datasetOpen Physics10.1515/phys-2024-006022:1Online publication date: 20-Jul-2024
  • Show More Cited By

Index Terms

  1. Frequency-aware Camouflaged Object Detection

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Transactions on Multimedia Computing, Communications, and Applications
    ACM Transactions on Multimedia Computing, Communications, and Applications  Volume 19, Issue 2
    March 2023
    540 pages
    ISSN:1551-6857
    EISSN:1551-6865
    DOI:10.1145/3572860
    • Editor:
    • Abdulmotaleb El Saddik
    Issue’s Table of Contents

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 23 March 2023
    Online AM: 30 June 2022
    Accepted: 23 June 2022
    Revised: 24 May 2022
    Received: 24 September 2021
    Published in TOMM Volume 19, Issue 2

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Camouflaged object detection
    2. frequency learning

    Qualifiers

    • Research-article

    Funding Sources

    • GRF
    • Research Grants Council of Hong Kong
    • Postgraduate Studentship (Mainland Schemes) from City University of Hong Kong

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)441
    • Downloads (Last 6 weeks)21
    Reflects downloads up to 01 Feb 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2025)Frequency-Guided Spatial Adaptation for Camouflaged Object DetectionIEEE Transactions on Multimedia10.1109/TMM.2024.352168127(72-83)Online publication date: 2025
    • (2025)CoNet: A Consistency-Oriented Network for Camouflaged Object SegmentationIEEE Transactions on Circuits and Systems for Video Technology10.1109/TCSVT.2024.346246535:1(287-299)Online publication date: Jan-2025
    • (2024)A new benchmark for camouflaged object detection: RGB-D camouflaged object detection datasetOpen Physics10.1515/phys-2024-006022:1Online publication date: 20-Jul-2024
    • (2024)Toward Oriented Fisheye Object Detection: Dataset and BaselineACM Transactions on Multimedia Computing, Communications, and Applications10.1145/370264021:1(1-19)Online publication date: 2-Nov-2024
    • (2024)TFSemantic: A Time–Frequency Semantic GAN Framework for Imbalanced Classification Using Radio SignalsACM Transactions on Sensor Networks10.1145/361409620:4(1-22)Online publication date: 11-May-2024
    • (2024)FINet: Frequency Injection Network for Lightweight Camouflaged Object DetectionIEEE Signal Processing Letters10.1109/LSP.2024.335641631(526-530)Online publication date: 2024
    • (2024)Weighted Dense Semantic Aggregation and Explicit Boundary Modeling for Camouflaged Object DetectionIEEE Sensors Journal10.1109/JSEN.2024.340172224:13(21108-21122)Online publication date: 1-Jul-2024
    • (2024)Camouflaged Object Detection using Multi-Level Feature Cross-Fusion2024 International Joint Conference on Neural Networks (IJCNN)10.1109/IJCNN60899.2024.10651348(1-8)Online publication date: 30-Jun-2024
    • (2024)TS-SAM: Two Small Steps for SAM, One Giant Leap for Abnormal detections2024 IEEE International Conference on Multimedia and Expo (ICME)10.1109/ICME57554.2024.10688209(1-6)Online publication date: 15-Jul-2024
    • (2024)A Novel Multiclass Object Detection Dataset Enriched With Frequency DataIEEE Access10.1109/ACCESS.2024.341616812(85551-85564)Online publication date: 2024
    • Show More Cited By

    View Options

    Login options

    Full Access

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    Full Text

    HTML Format

    View this article in HTML Format.

    HTML Format

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media