Deep Transformer Based Video Inpainting Using Fast Fourier Tokenization
Bridging distant space-time interactions is important for high-quality video inpainting with large moving masks. Most existing technologies exploit patch similarities within the frames, or leaverage large-scale training data to fill the hole along spatial and temporal dimensions. Recent works introd...
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2024-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/10418237/ |
_version_ | 1797311675166097408 |
---|---|
author | Taewan Kim Jinwoo Kim Heeseok Oh Jiwoo Kang |
author_facet | Taewan Kim Jinwoo Kim Heeseok Oh Jiwoo Kang |
author_sort | Taewan Kim |
collection | DOAJ |
description | Bridging distant space-time interactions is important for high-quality video inpainting with large moving masks. Most existing technologies exploit patch similarities within the frames, or leaverage large-scale training data to fill the hole along spatial and temporal dimensions. Recent works introduce promissing Transformer architecture into deep video inpainting to escape from the dominanace of nearby interactions and achieve superior performance than their baselines. However, such methods still struggle to complete larger holes containing complicated scenes. To alleviate this issue, we first employ a fast Fourier convolutions, which cover the frame-wide receptive field, for token representation. Then, the token passes through the seperated spatio-temporal transformer to explicitly moel the long-range context relations and simultaneously complete the missing regions in all input frames. By formulating video inpainting as a directionless sequence-to-sequence prediction task, our model fills visually consistent content, even under conditions such as large missing areas or complex geometries. Furthermore, our spatio-temporal transformer iteratively fills the hole from the boundary enabling it to exploit rich contextual information. We validate the superiority of the proposed model by using standard stationary masks and more realistic moving object masks. Both qualitative and quantitative results show that our model compares favorably against the state-of-the-art algorithms. |
first_indexed | 2024-03-08T02:04:10Z |
format | Article |
id | doaj.art-c86117bafb7f4ad1919228ed6483d1a4 |
institution | Directory Open Access Journal |
issn | 2169-3536 |
language | English |
last_indexed | 2024-03-08T02:04:10Z |
publishDate | 2024-01-01 |
publisher | IEEE |
record_format | Article |
series | IEEE Access |
spelling | doaj.art-c86117bafb7f4ad1919228ed6483d1a42024-02-14T00:01:23ZengIEEEIEEE Access2169-35362024-01-0112217232173610.1109/ACCESS.2024.336128310418237Deep Transformer Based Video Inpainting Using Fast Fourier TokenizationTaewan Kim0https://orcid.org/0000-0003-3319-7797Jinwoo Kim1Heeseok Oh2https://orcid.org/0000-0002-0920-7281Jiwoo Kang3https://orcid.org/0000-0002-0920-7281Data Science Major, Dongduk Women’s University, Seoul, South KoreaDepartment of Electrical and Electronic Engineering, Yonsei University, Seoul, South KoreaDepartment of Applied AI, Hansung University, Seoul, South KoreaDivision of Artificial Intelligence Engineering, Sookmyung Women’s University, Seoul, South KoreaBridging distant space-time interactions is important for high-quality video inpainting with large moving masks. Most existing technologies exploit patch similarities within the frames, or leaverage large-scale training data to fill the hole along spatial and temporal dimensions. Recent works introduce promissing Transformer architecture into deep video inpainting to escape from the dominanace of nearby interactions and achieve superior performance than their baselines. However, such methods still struggle to complete larger holes containing complicated scenes. To alleviate this issue, we first employ a fast Fourier convolutions, which cover the frame-wide receptive field, for token representation. Then, the token passes through the seperated spatio-temporal transformer to explicitly moel the long-range context relations and simultaneously complete the missing regions in all input frames. By formulating video inpainting as a directionless sequence-to-sequence prediction task, our model fills visually consistent content, even under conditions such as large missing areas or complex geometries. Furthermore, our spatio-temporal transformer iteratively fills the hole from the boundary enabling it to exploit rich contextual information. We validate the superiority of the proposed model by using standard stationary masks and more realistic moving object masks. Both qualitative and quantitative results show that our model compares favorably against the state-of-the-art algorithms.https://ieeexplore.ieee.org/document/10418237/Video inpaintingvideo completionfree-form inpaintingobject removaladversarial learning |
spellingShingle | Taewan Kim Jinwoo Kim Heeseok Oh Jiwoo Kang Deep Transformer Based Video Inpainting Using Fast Fourier Tokenization IEEE Access Video inpainting video completion free-form inpainting object removal adversarial learning |
title | Deep Transformer Based Video Inpainting Using Fast Fourier Tokenization |
title_full | Deep Transformer Based Video Inpainting Using Fast Fourier Tokenization |
title_fullStr | Deep Transformer Based Video Inpainting Using Fast Fourier Tokenization |
title_full_unstemmed | Deep Transformer Based Video Inpainting Using Fast Fourier Tokenization |
title_short | Deep Transformer Based Video Inpainting Using Fast Fourier Tokenization |
title_sort | deep transformer based video inpainting using fast fourier tokenization |
topic | Video inpainting video completion free-form inpainting object removal adversarial learning |
url | https://ieeexplore.ieee.org/document/10418237/ |
work_keys_str_mv | AT taewankim deeptransformerbasedvideoinpaintingusingfastfouriertokenization AT jinwookim deeptransformerbasedvideoinpaintingusingfastfouriertokenization AT heeseokoh deeptransformerbasedvideoinpaintingusingfastfouriertokenization AT jiwookang deeptransformerbasedvideoinpaintingusingfastfouriertokenization |