YOLOv5-Fog: A Multiobjective Visual Detection Algorithm for Fog Driving Scenes Based on Improved YOLOv5

With the rapid development of deep learning in recent years, the level of automatic driving perception has also increased substantially. However, automatic driving perception under adverse conditions, such as fog, remains a significant obstacle. The existing fog-oriented detection algorithms are una...

Full description

Saved in:
Bibliographic Details
Published inIEEE transactions on instrumentation and measurement Vol. 71; pp. 1 - 12
Main Authors Wang, Hai, Xu, Yansong, He, Youguo, Cai, Yingfeng, Chen, Long, Li, Yicheng, Sotelo, Miguel Angel, Li, Zhixiong
Format Journal Article
LanguageEnglish
Published New York IEEE 2022
The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
Subjects
Online AccessGet full text
ISSN0018-9456
1557-9662
DOI10.1109/TIM.2022.3196954

Cover

More Information
Summary:With the rapid development of deep learning in recent years, the level of automatic driving perception has also increased substantially. However, automatic driving perception under adverse conditions, such as fog, remains a significant obstacle. The existing fog-oriented detection algorithms are unable to simultaneously address the detection accuracy and detection speed. Based on improved YOLOv5, this work provides a multiobject detection network for fog driving scenes. We construct a synthetic fog dataset by using the dataset of a virtual scene and the depth information of the image. Second, we present a detection network for driving in fog based on improved YOLOv5. The ResNeXt model, which has been modified by structural re-parameterization, serves as the model's backbone. We build a new feature enhancement module (FEM) in response to the lack of features in fog scene images and use the attention mechanism to help the detection network pay more attention to the more useful features in the fog scenes. The test results show that the proposed fog multitarget detection network outperforms the original YOLOv5 in terms of detection accuracy and speed. The accuracy of the Real-world Task-driven Testing Set (RTTS) public dataset is 77.8%, and the detection speed is 31 frames/s, which is 14 frames faster as compared with the original YOLOv5.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ISSN:0018-9456
1557-9662
DOI:10.1109/TIM.2022.3196954