Journal of Jilin University (Information Science Edition) ›› 2024, Vol. 42 ›› Issue (3): 573-578.

Previous Articles    

 Research on Multi-Modal RGB-T Based Saliency Target Detection Algorithm

LIU Dong, BI Hongbo, REN Siqi, YU Xin, ZHANG Cong   

  1. School of Electrical and Information Engineering, Northeastern Petroleum University, Daqing 163318, China
  • Received:2023-04-13 Online:2024-06-18 Published:2024-06-18

Abstract: To address the problem that RGB ( Red Green Blue ) modal and thermal modal information representations are inconsistent in form and feature information can not be effectively mined and fused, a new joint attention reinforcement network-FCNet ( Feature Sharpening and Cross-modal Feature Fusion Net ) is proposed. Firstly, the image feature mapping capability is enhanced by a two-dimensional attention mechanism. Then, a cross-modal feature fusion mechanism is used to capture the target region. Finally, a layer-by-layer decoding structure is used to eliminate background interference and optimize the detection target. The experimental results demonstrate that the improved algorithm has fewer parameters and shorter operation times, and the overall detection performance of the model is better than that of existing multimodal detection models.

Key words: multimodality, RGB-Thermal(RGB-T), feature sharpening module, cross-modal fusion mechanism

CLC Number: 

  • TP391. 41