Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
loading
Papers Papers/2022 Papers Papers/2022

Research.Publish.Connect.

Paper

Authors: Mohamed Amine Marnissi 1 ; 2 ; 3 ; Ikram Hattab 2 ; 4 ; Hajer Fradi 4 ; 3 ; Anis Sahbani 2 and Najoua Essoukri Ben Amara 5 ; 3

Affiliations: 1 Ecole Nationale d’Ingénieurs de Sfax, Université de Sfax, 3038, Sfax, Tunisia ; 2 Enova Robotics, Novation City, 4000, Sousse, Tunisia ; 3 LATIS- Laboratory of Advanced Technology and Intelligent Systems, Université de Sousse, 4023, Sousse, Tunisia ; 4 Institut Supérieur des Sciences Appliquées et de Technologie, Université de Sousse, 4023, Sousse, Tunisia ; 5 Ecole Nationale d’Ingénieurs de Sousse, Université de Sousse, 4023, Sousse, Tunisia

Keyword(s): Deep Learning, Object Detection, YOLO, Visible and Thermal Cameras, Robotic Vision, Saliency Map, Transformer, Features Fusion.

Abstract: In this paper, we focus on the problem of automatic pedestrian detection for surveillance applications. Particularly, the main goal is to perform real-time detection from both visible and thermal cameras for complementary aspects. To handle that, a fusion network that uses features from both inputs and performs augmentation by means of visual saliency transformation is proposed. This fusion process is incorporated into YOLO-v3 as base architecture. The resulting detection model is trained in a paired setting in order to improve the results compared to the detection of each single input. To prove the effectiveness of the proposed fusion framework, several experiments are conducted on KAIST multi-spectral dataset. From the obtained results, it has been shown superior results compared to single inputs and to other fusion schemes. The proposed approach has also the advantage of a very low computational cost, which is quite important for real-time applications. To prove that, additional t ests on a security robot are presented as well. (More)

CC BY-NC-ND 4.0

Sign In Guest: Register as new SciTePress user now for free.

Sign In SciTePress user: please login.

PDF ImageMy Papers

You are not signed in, therefore limits apply to your IP address 70.40.220.129

In the current month:
Recent papers: 100 available of 100 total
2+ years older papers: 200 available of 200 total

Paper citation in several formats:
Marnissi, M.; Hattab, I.; Fradi, H.; Sahbani, A. and Ben Amara, N. (2022). Bispectral Pedestrian Detection Augmented with Saliency Maps using Transformer. In Proceedings of the 17th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP 2022) - Volume 5: VISAPP; ISBN 978-989-758-555-5; ISSN 2184-4321, SciTePress, pages 275-284. DOI: 10.5220/0010913000003124

@conference{visapp22,
author={Mohamed Amine Marnissi. and Ikram Hattab. and Hajer Fradi. and Anis Sahbani. and Najoua Essoukri {Ben Amara}.},
title={Bispectral Pedestrian Detection Augmented with Saliency Maps using Transformer},
booktitle={Proceedings of the 17th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP 2022) - Volume 5: VISAPP},
year={2022},
pages={275-284},
publisher={SciTePress},
organization={INSTICC},
doi={10.5220/0010913000003124},
isbn={978-989-758-555-5},
issn={2184-4321},
}

TY - CONF

JO - Proceedings of the 17th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP 2022) - Volume 5: VISAPP
TI - Bispectral Pedestrian Detection Augmented with Saliency Maps using Transformer
SN - 978-989-758-555-5
IS - 2184-4321
AU - Marnissi, M.
AU - Hattab, I.
AU - Fradi, H.
AU - Sahbani, A.
AU - Ben Amara, N.
PY - 2022
SP - 275
EP - 284
DO - 10.5220/0010913000003124
PB - SciTePress