Video understanding using multimodal deep learning
<p>Our experience of the world is multimodal, however deep learning networks have been traditionally designed for and trained on unimodal inputs such as images, audio segments or text. In this thesis we develop strategies to exploit multimodal information (in the form of vision, text, speech a...
Auteur principal: | Nagrani, A |
---|---|
Autres auteurs: | Zisserman, A |
Format: | Thèse |
Langue: | English |
Publié: |
2020
|
Sujets: |
Documents similaires
-
Sign language understanding using multimodal learning
par: Momeni, L
Publié: (2024) -
Understanding Multimodal Popularity Prediction of Social Media Videos With Self-Attention
par: Adam Bielski, et autres
Publié: (2018-01-01) -
End-to-end learning, and audio-visual human-centric video understanding
par: Brown, A
Publié: (2022) -
Holistic image understanding with deep learning and dense random fields
par: Zheng, S
Publié: (2016) -
Learning with multimodal self-supervision
par: Chen, H
Publié: (2021)