View-Invariant Spatiotemporal Attentive Motion Planning and Control Network for Autonomous Vehicles

Autonomous driving vehicles (ADVs) are sleeping giant intelligent machines that perceive their environment and make driving decisions. Most existing ADSs are built as hand-engineered perception-planning-control pipelines. However, designing generalized handcrafted rules for autonomous driving in an...

Full description

Bibliographic Details
Main Authors: Melese Ayalew, Shijie Zhou, Imran Memon, Md Belal Bin Heyat, Faijan Akhtar, Xiaojuan Zhang
Format: Article
Language:English
Published: MDPI AG 2022-12-01
Series:Machines
Subjects:
Online Access:https://www.mdpi.com/2075-1702/10/12/1193
_version_ 1827638026069606400
author Melese Ayalew
Shijie Zhou
Imran Memon
Md Belal Bin Heyat
Faijan Akhtar
Xiaojuan Zhang
author_facet Melese Ayalew
Shijie Zhou
Imran Memon
Md Belal Bin Heyat
Faijan Akhtar
Xiaojuan Zhang
author_sort Melese Ayalew
collection DOAJ
description Autonomous driving vehicles (ADVs) are sleeping giant intelligent machines that perceive their environment and make driving decisions. Most existing ADSs are built as hand-engineered perception-planning-control pipelines. However, designing generalized handcrafted rules for autonomous driving in an urban environment is complex. An alternative approach is imitation learning (IL) from human driving demonstrations. However, most previous studies on IL for autonomous driving face several critical challenges: (1) poor generalization ability toward the unseen environment due to distribution shift problems such as changes in driving views and weather conditions; (2) lack of interpretability; and (3) mostly trained to learn the single driving task. To address these challenges, we propose a view-invariant spatiotemporal attentive planning and control network for autonomous vehicles. The proposed method first extracts spatiotemporal representations from images of a front and top driving view sequence through attentive Siamese 3DResNet. Then, the maximum mean discrepancy loss (MMD) is employed to minimize spatiotemporal discrepancies between these driving views and produce an invariant spatiotemporal representation, which reduces domain shift due to view change. Finally, the multitasking learning (MTL) method is employed to jointly train trajectory planning and high-level control tasks based on learned representations and previous motions. Results of extensive experimental evaluations on a large autonomous driving dataset with various weather/lighting conditions verified that the proposed method is effective for feasible motion planning and control in autonomous vehicles.
first_indexed 2024-03-09T16:09:55Z
format Article
id doaj.art-4ff14249c1464851a57519076fd4d831
institution Directory Open Access Journal
issn 2075-1702
language English
last_indexed 2024-03-09T16:09:55Z
publishDate 2022-12-01
publisher MDPI AG
record_format Article
series Machines
spelling doaj.art-4ff14249c1464851a57519076fd4d8312023-11-24T16:17:12ZengMDPI AGMachines2075-17022022-12-011012119310.3390/machines10121193View-Invariant Spatiotemporal Attentive Motion Planning and Control Network for Autonomous VehiclesMelese Ayalew0Shijie Zhou1Imran Memon2Md Belal Bin Heyat3Faijan Akhtar4Xiaojuan Zhang5School of Information and Software Engineering, University of Electronic Science and Technology of China, Chengdu 610054, ChinaSchool of Information and Software Engineering, University of Electronic Science and Technology of China, Chengdu 610054, ChinaDepartment of Computer Science, Shahdadkot Campus, Shah Abdul Latif University, Khairpur 66111, PakistanIoT Research Center, College of Computer Science and Software Engineering, Shenzhen University, Shenzhen 518060, ChinaSchool of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 610054, ChinaZhejiang CloudNeedle Information Technology Co., Ltd., Hangzhou 311121, ChinaAutonomous driving vehicles (ADVs) are sleeping giant intelligent machines that perceive their environment and make driving decisions. Most existing ADSs are built as hand-engineered perception-planning-control pipelines. However, designing generalized handcrafted rules for autonomous driving in an urban environment is complex. An alternative approach is imitation learning (IL) from human driving demonstrations. However, most previous studies on IL for autonomous driving face several critical challenges: (1) poor generalization ability toward the unseen environment due to distribution shift problems such as changes in driving views and weather conditions; (2) lack of interpretability; and (3) mostly trained to learn the single driving task. To address these challenges, we propose a view-invariant spatiotemporal attentive planning and control network for autonomous vehicles. The proposed method first extracts spatiotemporal representations from images of a front and top driving view sequence through attentive Siamese 3DResNet. Then, the maximum mean discrepancy loss (MMD) is employed to minimize spatiotemporal discrepancies between these driving views and produce an invariant spatiotemporal representation, which reduces domain shift due to view change. Finally, the multitasking learning (MTL) method is employed to jointly train trajectory planning and high-level control tasks based on learned representations and previous motions. Results of extensive experimental evaluations on a large autonomous driving dataset with various weather/lighting conditions verified that the proposed method is effective for feasible motion planning and control in autonomous vehicles.https://www.mdpi.com/2075-1702/10/12/1193autonomous vehiclesdeep learninginvariant representation learningmotion planningspatiotemporal attentionvehicle control
spellingShingle Melese Ayalew
Shijie Zhou
Imran Memon
Md Belal Bin Heyat
Faijan Akhtar
Xiaojuan Zhang
View-Invariant Spatiotemporal Attentive Motion Planning and Control Network for Autonomous Vehicles
Machines
autonomous vehicles
deep learning
invariant representation learning
motion planning
spatiotemporal attention
vehicle control
title View-Invariant Spatiotemporal Attentive Motion Planning and Control Network for Autonomous Vehicles
title_full View-Invariant Spatiotemporal Attentive Motion Planning and Control Network for Autonomous Vehicles
title_fullStr View-Invariant Spatiotemporal Attentive Motion Planning and Control Network for Autonomous Vehicles
title_full_unstemmed View-Invariant Spatiotemporal Attentive Motion Planning and Control Network for Autonomous Vehicles
title_short View-Invariant Spatiotemporal Attentive Motion Planning and Control Network for Autonomous Vehicles
title_sort view invariant spatiotemporal attentive motion planning and control network for autonomous vehicles
topic autonomous vehicles
deep learning
invariant representation learning
motion planning
spatiotemporal attention
vehicle control
url https://www.mdpi.com/2075-1702/10/12/1193
work_keys_str_mv AT meleseayalew viewinvariantspatiotemporalattentivemotionplanningandcontrolnetworkforautonomousvehicles
AT shijiezhou viewinvariantspatiotemporalattentivemotionplanningandcontrolnetworkforautonomousvehicles
AT imranmemon viewinvariantspatiotemporalattentivemotionplanningandcontrolnetworkforautonomousvehicles
AT mdbelalbinheyat viewinvariantspatiotemporalattentivemotionplanningandcontrolnetworkforautonomousvehicles
AT faijanakhtar viewinvariantspatiotemporalattentivemotionplanningandcontrolnetworkforautonomousvehicles
AT xiaojuanzhang viewinvariantspatiotemporalattentivemotionplanningandcontrolnetworkforautonomousvehicles