Occlusion Avoidance in a Simulated Environment Using Reinforcement Learning
Neural network-based solutions have revolutionized the field of computer vision by achieving outstanding performance in a number of applications. Yet, while these deep learning models outclass previous methods, they still have significant shortcomings relating to generalization and robustness to inp...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2023-02-01
|
Series: | Applied Sciences |
Subjects: | |
Online Access: | https://www.mdpi.com/2076-3417/13/5/3090 |
_version_ | 1797615757398376448 |
---|---|
author | Márton Szemenyei Mátyás Szántó |
author_facet | Márton Szemenyei Mátyás Szántó |
author_sort | Márton Szemenyei |
collection | DOAJ |
description | Neural network-based solutions have revolutionized the field of computer vision by achieving outstanding performance in a number of applications. Yet, while these deep learning models outclass previous methods, they still have significant shortcomings relating to generalization and robustness to input disturbances, such as occlusion. Most existing methods that tackle this latter problem use passive neural network architectures that are unable to act on and, thus, influence the observed scene. In this paper, we argue that an active observer agent may be able to achieve superior performance by changing the parameters of the scene, thus, avoiding occlusion by moving to a different position in the scene. To demonstrate this, a reinforcement learning environment is introduced that implements OpenAI Gym’s interface, and allows the creation of synthetic scenes with realistic occlusion. The environment is implemented using differentiable rendering, allowing us to perform direct gradient-based optimization of the camera position. Moreover, two additional methods are also presented, one utilizing self-supervised learning to predict occlusion segments, and optimal camera positions, while the other learns to avoid occlusion using Reinforcement Learning. We present comparative experiments of the proposed methods to demonstrate their efficiency. It was shown, via Bayesian <i>t</i>-tests, that the neural network-based methods credibly outperformed the gradient-based avoidance strategy by avoiding occlusion with an average of 5.0 fewer steps in multi-object scenes. |
first_indexed | 2024-03-11T07:31:17Z |
format | Article |
id | doaj.art-e9bc3817a40048c4a3a7925980a31c0d |
institution | Directory Open Access Journal |
issn | 2076-3417 |
language | English |
last_indexed | 2024-03-11T07:31:17Z |
publishDate | 2023-02-01 |
publisher | MDPI AG |
record_format | Article |
series | Applied Sciences |
spelling | doaj.art-e9bc3817a40048c4a3a7925980a31c0d2023-11-17T07:19:07ZengMDPI AGApplied Sciences2076-34172023-02-01135309010.3390/app13053090Occlusion Avoidance in a Simulated Environment Using Reinforcement LearningMárton Szemenyei0Mátyás Szántó1Department of Control Engineering and Information Technology, Budapest University of Technology and Economics, Műegyetem rkp. 3., H-1111 Budapest, HungaryDepartment of Control Engineering and Information Technology, Budapest University of Technology and Economics, Műegyetem rkp. 3., H-1111 Budapest, HungaryNeural network-based solutions have revolutionized the field of computer vision by achieving outstanding performance in a number of applications. Yet, while these deep learning models outclass previous methods, they still have significant shortcomings relating to generalization and robustness to input disturbances, such as occlusion. Most existing methods that tackle this latter problem use passive neural network architectures that are unable to act on and, thus, influence the observed scene. In this paper, we argue that an active observer agent may be able to achieve superior performance by changing the parameters of the scene, thus, avoiding occlusion by moving to a different position in the scene. To demonstrate this, a reinforcement learning environment is introduced that implements OpenAI Gym’s interface, and allows the creation of synthetic scenes with realistic occlusion. The environment is implemented using differentiable rendering, allowing us to perform direct gradient-based optimization of the camera position. Moreover, two additional methods are also presented, one utilizing self-supervised learning to predict occlusion segments, and optimal camera positions, while the other learns to avoid occlusion using Reinforcement Learning. We present comparative experiments of the proposed methods to demonstrate their efficiency. It was shown, via Bayesian <i>t</i>-tests, that the neural network-based methods credibly outperformed the gradient-based avoidance strategy by avoiding occlusion with an average of 5.0 fewer steps in multi-object scenes.https://www.mdpi.com/2076-3417/13/5/3090computer visionobject detectiondifferentiable renderingself-supervised learningneural networksreinforcement learning |
spellingShingle | Márton Szemenyei Mátyás Szántó Occlusion Avoidance in a Simulated Environment Using Reinforcement Learning Applied Sciences computer vision object detection differentiable rendering self-supervised learning neural networks reinforcement learning |
title | Occlusion Avoidance in a Simulated Environment Using Reinforcement Learning |
title_full | Occlusion Avoidance in a Simulated Environment Using Reinforcement Learning |
title_fullStr | Occlusion Avoidance in a Simulated Environment Using Reinforcement Learning |
title_full_unstemmed | Occlusion Avoidance in a Simulated Environment Using Reinforcement Learning |
title_short | Occlusion Avoidance in a Simulated Environment Using Reinforcement Learning |
title_sort | occlusion avoidance in a simulated environment using reinforcement learning |
topic | computer vision object detection differentiable rendering self-supervised learning neural networks reinforcement learning |
url | https://www.mdpi.com/2076-3417/13/5/3090 |
work_keys_str_mv | AT martonszemenyei occlusionavoidanceinasimulatedenvironmentusingreinforcementlearning AT matyasszanto occlusionavoidanceinasimulatedenvironmentusingreinforcementlearning |