A pedestrian detection method based on YoloV5 RGB and thermal image data fusion is proposed in this paper. This method uses two yolov5 branch structures to learn the features of RGB and thermal image data respectively, and finally uses multimodal features for fusion detection. It includes the following stages. Firstly, we use yolov5 network as a branch network to learn features from paired RGB and thermal image data. The two yoloV5-based backbone networks extract the features of the two modes for preliminary fusion, and then obtain the importance parameters of the two modes through feature compression and extraction. The effective information is enhanced, and redundant information is eliminated by multiplying the initial features. Finally, the fusion feature is used for target detection. Through this method, the detection effect is improved. We have done a lot of experiments on the public KAIST and VOT2019 pedestrian data set, and the experiments show that our method is better than the advanced method.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.