Read and attend: temporal localisation in sign language videos
The objective of this work is to annotate sign instances across a broad vocabulary in continuous sign language. We train a Transformer model to ingest a continuous signing stream and output a sequence of written tokens on a large-scale collection of signing footage with weakly-aligned subtitles. We...
Main Authors: | Varol, G, Momeni, L, Albanie, S, Afouras, T, Zisserman, A |
---|---|
Format: | Conference item |
Language: | English |
Published: |
IEEE
2021
|
Similar Items
-
Aligning subtitles in sign language videos
by: Bull, H, et al.
Published: (2022) -
Scaling up sign spotting through sign language dictionaries
by: Varol, G, et al.
Published: (2022) -
Watch, read and lookup: learning to spot signs from multiple supervisors
by: Momeni, L, et al.
Published: (2021) -
Automatic dense annotation of large-vocabulary sign language videos
by: Momeni, L, et al.
Published: (2022) -
Weakly-supervised fingerspelling recognition in British Sign Language videos
by: Prajwal, KR, et al.
Published: (2022)