ST-MAE: robust lane detection in continuous multi-frame driving scenes based on a deep hybrid network

Abstract Lane detection is one of the key techniques to realize advanced driving assistance and automatic driving. However, lane detection networks based on deep learning have significant shortcomings. The detection results are often unsatisfactory when there are shadows, degraded lane markings, and...

Full description

Bibliographic Details
Main Authors: Rongyun Zhang, Yufeng Du, Peicheng Shi, Lifeng Zhao, Yaming Liu, Haoran Li
Format: Article
Language:English
Published: Springer 2022-11-01
Series:Complex & Intelligent Systems
Subjects:
Online Access:https://doi.org/10.1007/s40747-022-00909-0
_version_ 1797674956059836416
author Rongyun Zhang
Yufeng Du
Peicheng Shi
Lifeng Zhao
Yaming Liu
Haoran Li
author_facet Rongyun Zhang
Yufeng Du
Peicheng Shi
Lifeng Zhao
Yaming Liu
Haoran Li
author_sort Rongyun Zhang
collection DOAJ
description Abstract Lane detection is one of the key techniques to realize advanced driving assistance and automatic driving. However, lane detection networks based on deep learning have significant shortcomings. The detection results are often unsatisfactory when there are shadows, degraded lane markings, and vehicle occlusion lanes. Therefore, a continuous multi-frame image sequence lane detection network is proposed. Specifically, the continuous six-frame image sequence is input into the network, in which the scene information of each frame image is extracted by an encoder composed of Swin Transformer blocks and input into the PredRNN. Continuous multi-frame of the driving scene is modeled as time-series by ST-LSTM blocks, and then, the shape changes and motion trajectory in the spatiotemporal sequence are effectively modeled. Finally, through the decoder composed of Swin Transformer blocks, the features are obtained and reconstructed to complete the detection task. Extensive experiments on two large-scale datasets demonstrate that the proposed method outperforms the competing methods in lane detection, especially in handling difficult situations. Experiments are carried out based on the TuSimple dataset. The results show: for easy scenes, the validation accuracy is 97.46%, the test accuracy is 97.37%, and the precision is 0.865. For complex scenes, the validation accuracy is 97.38%, the test accuracy is 97.29%, and the precision is 0.859. The running time is 4.4 ms. Experiments are carried out based on the CULane dataset. The results show that, for easy scenes, the validation accuracy is 97.03%, the test accuracy is 96.84%, and the precision is 0.837. For complex scenes, the validation accuracy is 96.18%, the test accuracy is 95.92%, and the precision is 0.829. The running time is 6.5 ms.
first_indexed 2024-03-11T22:07:40Z
format Article
id doaj.art-2d94c6a5b779466e849e76c01336eb0e
institution Directory Open Access Journal
issn 2199-4536
2198-6053
language English
last_indexed 2024-03-11T22:07:40Z
publishDate 2022-11-01
publisher Springer
record_format Article
series Complex & Intelligent Systems
spelling doaj.art-2d94c6a5b779466e849e76c01336eb0e2023-09-24T11:35:22ZengSpringerComplex & Intelligent Systems2199-45362198-60532022-11-01954837485510.1007/s40747-022-00909-0ST-MAE: robust lane detection in continuous multi-frame driving scenes based on a deep hybrid networkRongyun Zhang0Yufeng Du1Peicheng Shi2Lifeng Zhao3Yaming Liu4Haoran Li5The School of Mechanical and Automotive Engineering, Anhui Polytechnic UniversityThe School of Mechanical and Automotive Engineering, Anhui Polytechnic UniversityAutomotive New Technology Anhui Engineering and Technology Research Center, Anhui Polytechnic UniversityThe School of Automotive and Traffic Engineering, Hefei University of TechnologyThe School of Mechanical and Automotive Engineering, Anhui Polytechnic UniversityThe School of Mechanical and Automotive Engineering, Anhui Polytechnic UniversityAbstract Lane detection is one of the key techniques to realize advanced driving assistance and automatic driving. However, lane detection networks based on deep learning have significant shortcomings. The detection results are often unsatisfactory when there are shadows, degraded lane markings, and vehicle occlusion lanes. Therefore, a continuous multi-frame image sequence lane detection network is proposed. Specifically, the continuous six-frame image sequence is input into the network, in which the scene information of each frame image is extracted by an encoder composed of Swin Transformer blocks and input into the PredRNN. Continuous multi-frame of the driving scene is modeled as time-series by ST-LSTM blocks, and then, the shape changes and motion trajectory in the spatiotemporal sequence are effectively modeled. Finally, through the decoder composed of Swin Transformer blocks, the features are obtained and reconstructed to complete the detection task. Extensive experiments on two large-scale datasets demonstrate that the proposed method outperforms the competing methods in lane detection, especially in handling difficult situations. Experiments are carried out based on the TuSimple dataset. The results show: for easy scenes, the validation accuracy is 97.46%, the test accuracy is 97.37%, and the precision is 0.865. For complex scenes, the validation accuracy is 97.38%, the test accuracy is 97.29%, and the precision is 0.859. The running time is 4.4 ms. Experiments are carried out based on the CULane dataset. The results show that, for easy scenes, the validation accuracy is 97.03%, the test accuracy is 96.84%, and the precision is 0.837. For complex scenes, the validation accuracy is 96.18%, the test accuracy is 95.92%, and the precision is 0.829. The running time is 6.5 ms.https://doi.org/10.1007/s40747-022-00909-0Lane detectionDeep learningSwin transformerContinuous multi-frame
spellingShingle Rongyun Zhang
Yufeng Du
Peicheng Shi
Lifeng Zhao
Yaming Liu
Haoran Li
ST-MAE: robust lane detection in continuous multi-frame driving scenes based on a deep hybrid network
Complex & Intelligent Systems
Lane detection
Deep learning
Swin transformer
Continuous multi-frame
title ST-MAE: robust lane detection in continuous multi-frame driving scenes based on a deep hybrid network
title_full ST-MAE: robust lane detection in continuous multi-frame driving scenes based on a deep hybrid network
title_fullStr ST-MAE: robust lane detection in continuous multi-frame driving scenes based on a deep hybrid network
title_full_unstemmed ST-MAE: robust lane detection in continuous multi-frame driving scenes based on a deep hybrid network
title_short ST-MAE: robust lane detection in continuous multi-frame driving scenes based on a deep hybrid network
title_sort st mae robust lane detection in continuous multi frame driving scenes based on a deep hybrid network
topic Lane detection
Deep learning
Swin transformer
Continuous multi-frame
url https://doi.org/10.1007/s40747-022-00909-0
work_keys_str_mv AT rongyunzhang stmaerobustlanedetectionincontinuousmultiframedrivingscenesbasedonadeephybridnetwork
AT yufengdu stmaerobustlanedetectionincontinuousmultiframedrivingscenesbasedonadeephybridnetwork
AT peichengshi stmaerobustlanedetectionincontinuousmultiframedrivingscenesbasedonadeephybridnetwork
AT lifengzhao stmaerobustlanedetectionincontinuousmultiframedrivingscenesbasedonadeephybridnetwork
AT yamingliu stmaerobustlanedetectionincontinuousmultiframedrivingscenesbasedonadeephybridnetwork
AT haoranli stmaerobustlanedetectionincontinuousmultiframedrivingscenesbasedonadeephybridnetwork