ViT VO - A Visual Odometry technique Using CNN-Transformer Hybrid Architecture
Localization is one of the main tasks involved in the operation of autonomous agents (e.g., vehicle, robot etc.). It allows them to be able to track their paths and properly detect and avoid obstacles. Visual Odometry (VO) is one of the techniques used for agent localization. VO involves estimating...
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
EDP Sciences
2023-01-01
|
Series: | ITM Web of Conferences |
Subjects: | |
Online Access: | https://www.itm-conferences.org/articles/itmconf/pdf/2023/04/itmconf_I3cs2023_01004.pdf |
_version_ | 1797775772992143360 |
---|---|
author | B Jayaraj P. J Ebin R Karthik P N Pournami |
author_facet | B Jayaraj P. J Ebin R Karthik P N Pournami |
author_sort | B Jayaraj P. |
collection | DOAJ |
description | Localization is one of the main tasks involved in the operation of autonomous agents (e.g., vehicle, robot etc.). It allows them to be able to track their paths and properly detect and avoid obstacles. Visual Odometry (VO) is one of the techniques used for agent localization. VO involves estimating the motion of an agent using the images taken by cameras attached to it. Conventional VO algorithms require specific workarounds for challenges posed by the working environment and the captured sensor data. On the other hand, Deep Learning approaches have shown tremendous efficiency and accuracy in tasks that require high degree of adaptability and scalability. In this work, a novel deep learning model is proposed to perform VO tasks for space robotic applications. The model consists of an optical flow estimation module which abstracts away scene-specific details from the input video sequence and produces an intermediate representation. The CNN module which follows next learn relative poses from the optical flow estimates. The final module is a state-of-the-art Vision Transformer, which learn absolute pose from the relative pose learnt by the CNN module. The model is trained on the KITTI dataset and has obtained a promising accuracy of approximately 2%. It has outperformed the baseline model, MagicVO, in a few sequences in the dataset. |
first_indexed | 2024-03-12T22:40:27Z |
format | Article |
id | doaj.art-92d2cfcbd6ce48af8ceefc4e7c48db5b |
institution | Directory Open Access Journal |
issn | 2271-2097 |
language | English |
last_indexed | 2024-03-12T22:40:27Z |
publishDate | 2023-01-01 |
publisher | EDP Sciences |
record_format | Article |
series | ITM Web of Conferences |
spelling | doaj.art-92d2cfcbd6ce48af8ceefc4e7c48db5b2023-07-21T09:41:32ZengEDP SciencesITM Web of Conferences2271-20972023-01-01540100410.1051/itmconf/20235401004itmconf_I3cs2023_01004ViT VO - A Visual Odometry technique Using CNN-Transformer Hybrid ArchitectureB Jayaraj P.0J Ebin1R Karthik2P N Pournami3National Insitute of Technology CalicutNational Insitute of Technology CalicutSED/ISG, Advanced Inertial Systems, ISRO Inertial Systems UnitNational Insitute of Technology CalicutLocalization is one of the main tasks involved in the operation of autonomous agents (e.g., vehicle, robot etc.). It allows them to be able to track their paths and properly detect and avoid obstacles. Visual Odometry (VO) is one of the techniques used for agent localization. VO involves estimating the motion of an agent using the images taken by cameras attached to it. Conventional VO algorithms require specific workarounds for challenges posed by the working environment and the captured sensor data. On the other hand, Deep Learning approaches have shown tremendous efficiency and accuracy in tasks that require high degree of adaptability and scalability. In this work, a novel deep learning model is proposed to perform VO tasks for space robotic applications. The model consists of an optical flow estimation module which abstracts away scene-specific details from the input video sequence and produces an intermediate representation. The CNN module which follows next learn relative poses from the optical flow estimates. The final module is a state-of-the-art Vision Transformer, which learn absolute pose from the relative pose learnt by the CNN module. The model is trained on the KITTI dataset and has obtained a promising accuracy of approximately 2%. It has outperformed the baseline model, MagicVO, in a few sequences in the dataset.https://www.itm-conferences.org/articles/itmconf/pdf/2023/04/itmconf_I3cs2023_01004.pdfvisual odometrydeep learningoptical flowconvolutional neural networksgenerative adversarial networkssequence-based models |
spellingShingle | B Jayaraj P. J Ebin R Karthik P N Pournami ViT VO - A Visual Odometry technique Using CNN-Transformer Hybrid Architecture ITM Web of Conferences visual odometry deep learning optical flow convolutional neural networks generative adversarial networks sequence-based models |
title | ViT VO - A Visual Odometry technique Using CNN-Transformer Hybrid Architecture |
title_full | ViT VO - A Visual Odometry technique Using CNN-Transformer Hybrid Architecture |
title_fullStr | ViT VO - A Visual Odometry technique Using CNN-Transformer Hybrid Architecture |
title_full_unstemmed | ViT VO - A Visual Odometry technique Using CNN-Transformer Hybrid Architecture |
title_short | ViT VO - A Visual Odometry technique Using CNN-Transformer Hybrid Architecture |
title_sort | vit vo a visual odometry technique using cnn transformer hybrid architecture |
topic | visual odometry deep learning optical flow convolutional neural networks generative adversarial networks sequence-based models |
url | https://www.itm-conferences.org/articles/itmconf/pdf/2023/04/itmconf_I3cs2023_01004.pdf |
work_keys_str_mv | AT bjayarajp vitvoavisualodometrytechniqueusingcnntransformerhybridarchitecture AT jebin vitvoavisualodometrytechniqueusingcnntransformerhybridarchitecture AT rkarthik vitvoavisualodometrytechniqueusingcnntransformerhybridarchitecture AT pnpournami vitvoavisualodometrytechniqueusingcnntransformerhybridarchitecture |