Occlusion Avoidance in a Simulated Environment Using Reinforcement Learning

Neural network-based solutions have revolutionized the field of computer vision by achieving outstanding performance in a number of applications. Yet, while these deep learning models outclass previous methods, they still have significant shortcomings relating to generalization and robustness to inp...

Full description

Bibliographic Details
Main Authors: Márton Szemenyei, Mátyás Szántó
Format: Article
Language:English
Published: MDPI AG 2023-02-01
Series:Applied Sciences
Subjects:
Online Access:https://www.mdpi.com/2076-3417/13/5/3090
_version_ 1797615757398376448
author Márton Szemenyei
Mátyás Szántó
author_facet Márton Szemenyei
Mátyás Szántó
author_sort Márton Szemenyei
collection DOAJ
description Neural network-based solutions have revolutionized the field of computer vision by achieving outstanding performance in a number of applications. Yet, while these deep learning models outclass previous methods, they still have significant shortcomings relating to generalization and robustness to input disturbances, such as occlusion. Most existing methods that tackle this latter problem use passive neural network architectures that are unable to act on and, thus, influence the observed scene. In this paper, we argue that an active observer agent may be able to achieve superior performance by changing the parameters of the scene, thus, avoiding occlusion by moving to a different position in the scene. To demonstrate this, a reinforcement learning environment is introduced that implements OpenAI Gym’s interface, and allows the creation of synthetic scenes with realistic occlusion. The environment is implemented using differentiable rendering, allowing us to perform direct gradient-based optimization of the camera position. Moreover, two additional methods are also presented, one utilizing self-supervised learning to predict occlusion segments, and optimal camera positions, while the other learns to avoid occlusion using Reinforcement Learning. We present comparative experiments of the proposed methods to demonstrate their efficiency. It was shown, via Bayesian <i>t</i>-tests, that the neural network-based methods credibly outperformed the gradient-based avoidance strategy by avoiding occlusion with an average of 5.0 fewer steps in multi-object scenes.
first_indexed 2024-03-11T07:31:17Z
format Article
id doaj.art-e9bc3817a40048c4a3a7925980a31c0d
institution Directory Open Access Journal
issn 2076-3417
language English
last_indexed 2024-03-11T07:31:17Z
publishDate 2023-02-01
publisher MDPI AG
record_format Article
series Applied Sciences
spelling doaj.art-e9bc3817a40048c4a3a7925980a31c0d2023-11-17T07:19:07ZengMDPI AGApplied Sciences2076-34172023-02-01135309010.3390/app13053090Occlusion Avoidance in a Simulated Environment Using Reinforcement LearningMárton Szemenyei0Mátyás Szántó1Department of Control Engineering and Information Technology, Budapest University of Technology and Economics, Műegyetem rkp. 3., H-1111 Budapest, HungaryDepartment of Control Engineering and Information Technology, Budapest University of Technology and Economics, Műegyetem rkp. 3., H-1111 Budapest, HungaryNeural network-based solutions have revolutionized the field of computer vision by achieving outstanding performance in a number of applications. Yet, while these deep learning models outclass previous methods, they still have significant shortcomings relating to generalization and robustness to input disturbances, such as occlusion. Most existing methods that tackle this latter problem use passive neural network architectures that are unable to act on and, thus, influence the observed scene. In this paper, we argue that an active observer agent may be able to achieve superior performance by changing the parameters of the scene, thus, avoiding occlusion by moving to a different position in the scene. To demonstrate this, a reinforcement learning environment is introduced that implements OpenAI Gym’s interface, and allows the creation of synthetic scenes with realistic occlusion. The environment is implemented using differentiable rendering, allowing us to perform direct gradient-based optimization of the camera position. Moreover, two additional methods are also presented, one utilizing self-supervised learning to predict occlusion segments, and optimal camera positions, while the other learns to avoid occlusion using Reinforcement Learning. We present comparative experiments of the proposed methods to demonstrate their efficiency. It was shown, via Bayesian <i>t</i>-tests, that the neural network-based methods credibly outperformed the gradient-based avoidance strategy by avoiding occlusion with an average of 5.0 fewer steps in multi-object scenes.https://www.mdpi.com/2076-3417/13/5/3090computer visionobject detectiondifferentiable renderingself-supervised learningneural networksreinforcement learning
spellingShingle Márton Szemenyei
Mátyás Szántó
Occlusion Avoidance in a Simulated Environment Using Reinforcement Learning
Applied Sciences
computer vision
object detection
differentiable rendering
self-supervised learning
neural networks
reinforcement learning
title Occlusion Avoidance in a Simulated Environment Using Reinforcement Learning
title_full Occlusion Avoidance in a Simulated Environment Using Reinforcement Learning
title_fullStr Occlusion Avoidance in a Simulated Environment Using Reinforcement Learning
title_full_unstemmed Occlusion Avoidance in a Simulated Environment Using Reinforcement Learning
title_short Occlusion Avoidance in a Simulated Environment Using Reinforcement Learning
title_sort occlusion avoidance in a simulated environment using reinforcement learning
topic computer vision
object detection
differentiable rendering
self-supervised learning
neural networks
reinforcement learning
url https://www.mdpi.com/2076-3417/13/5/3090
work_keys_str_mv AT martonszemenyei occlusionavoidanceinasimulatedenvironmentusingreinforcementlearning
AT matyasszanto occlusionavoidanceinasimulatedenvironmentusingreinforcementlearning