YOLOv5-Fog: A Multiobjective Visual Detection Algorithm for Fog Driving Scenes Based on Improved YOLOv5
With the rapid development of deep learning in recent years, the level of automatic driving perception has also increased substantially. However, automatic driving perception under adverse conditions, such as fog, remains a significant obstacle. The existing fog-oriented detection algorithms are una...
        Saved in:
      
    
          | Published in | IEEE transactions on instrumentation and measurement Vol. 71; pp. 1 - 12 | 
|---|---|
| Main Authors | , , , , , , , | 
| Format | Journal Article | 
| Language | English | 
| Published | 
        New York
          IEEE
    
        2022
     The Institute of Electrical and Electronics Engineers, Inc. (IEEE)  | 
| Subjects | |
| Online Access | Get full text | 
| ISSN | 0018-9456 1557-9662  | 
| DOI | 10.1109/TIM.2022.3196954 | 
Cover
| Summary: | With the rapid development of deep learning in recent years, the level of automatic driving perception has also increased substantially. However, automatic driving perception under adverse conditions, such as fog, remains a significant obstacle. The existing fog-oriented detection algorithms are unable to simultaneously address the detection accuracy and detection speed. Based on improved YOLOv5, this work provides a multiobject detection network for fog driving scenes. We construct a synthetic fog dataset by using the dataset of a virtual scene and the depth information of the image. Second, we present a detection network for driving in fog based on improved YOLOv5. The ResNeXt model, which has been modified by structural re-parameterization, serves as the model's backbone. We build a new feature enhancement module (FEM) in response to the lack of features in fog scene images and use the attention mechanism to help the detection network pay more attention to the more useful features in the fog scenes. The test results show that the proposed fog multitarget detection network outperforms the original YOLOv5 in terms of detection accuracy and speed. The accuracy of the Real-world Task-driven Testing Set (RTTS) public dataset is 77.8%, and the detection speed is 31 frames/s, which is 14 frames faster as compared with the original YOLOv5. | 
|---|---|
| Bibliography: | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 14  | 
| ISSN: | 0018-9456 1557-9662  | 
| DOI: | 10.1109/TIM.2022.3196954 |