Unsupervised Phonetic Category Learning from Audio and Visual Input
Understanding how children learn the phonetic categories of their native language is an open area of research in cognitive science and child language development. However, despite experimental evidence that phonetic processing is very often a multimodal phenomenon (involving both auditory and visual...
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis |
Published: |
Massachusetts Institute of Technology
2023
|
Online Access: | https://hdl.handle.net/1721.1/151659 |
_version_ | 1826213014181773312 |
---|---|
author | Zhi, Sophia |
author2 | Levy, Roger |
author_facet | Levy, Roger Zhi, Sophia |
author_sort | Zhi, Sophia |
collection | MIT |
description | Understanding how children learn the phonetic categories of their native language is an open area of research in cognitive science and child language development. However, despite experimental evidence that phonetic processing is very often a multimodal phenomenon (involving both auditory and visual cues), computational research has primarily modeled phonetic category learning as a function of only auditory input. In this thesis, I investigate whether multimodal information benefits phonetic category learning under a clustering model. Due to the lack of an appropriate dataset, I also introduce a method for creating a high-quality dataset of synthetic videos of speakers’ faces for an existing audio corpus. This model trained and tested on audiovisual data achieves up to a 9.1% improvement on a phoneme discrimination battery over the random baseline compared to a model trained and tested on only audio data. The audiovisual model also outperforms the audio model by up to 4.7% over the baseline when both are tested on audio-only data, suggesting that visual information guides the learner towards better clusters. Further analysis indicates that visual information benefits most, but not all, phonemic contrasts. In follow-up analyses, I investigate the learned audiovisual clusters and their relationship to auditory gestures and phones, finding that the clusters capture a unit of speech smaller than phonemes. This work demonstrates the benefit of visual information to a computational model of phonetic category learning, suggesting that children may benefit substantively by using visual cues while learning phonetic categories. |
first_indexed | 2024-09-23T15:41:58Z |
format | Thesis |
id | mit-1721.1/151659 |
institution | Massachusetts Institute of Technology |
last_indexed | 2024-09-23T15:41:58Z |
publishDate | 2023 |
publisher | Massachusetts Institute of Technology |
record_format | dspace |
spelling | mit-1721.1/1516592023-08-01T03:10:17Z Unsupervised Phonetic Category Learning from Audio and Visual Input Zhi, Sophia Levy, Roger Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science Understanding how children learn the phonetic categories of their native language is an open area of research in cognitive science and child language development. However, despite experimental evidence that phonetic processing is very often a multimodal phenomenon (involving both auditory and visual cues), computational research has primarily modeled phonetic category learning as a function of only auditory input. In this thesis, I investigate whether multimodal information benefits phonetic category learning under a clustering model. Due to the lack of an appropriate dataset, I also introduce a method for creating a high-quality dataset of synthetic videos of speakers’ faces for an existing audio corpus. This model trained and tested on audiovisual data achieves up to a 9.1% improvement on a phoneme discrimination battery over the random baseline compared to a model trained and tested on only audio data. The audiovisual model also outperforms the audio model by up to 4.7% over the baseline when both are tested on audio-only data, suggesting that visual information guides the learner towards better clusters. Further analysis indicates that visual information benefits most, but not all, phonemic contrasts. In follow-up analyses, I investigate the learned audiovisual clusters and their relationship to auditory gestures and phones, finding that the clusters capture a unit of speech smaller than phonemes. This work demonstrates the benefit of visual information to a computational model of phonetic category learning, suggesting that children may benefit substantively by using visual cues while learning phonetic categories. M.Eng. 2023-07-31T19:57:03Z 2023-07-31T19:57:03Z 2023-06 2023-06-06T16:35:03.470Z Thesis https://hdl.handle.net/1721.1/151659 In Copyright - Educational Use Permitted Copyright retained by author(s) https://rightsstatements.org/page/InC-EDU/1.0/ application/pdf Massachusetts Institute of Technology |
spellingShingle | Zhi, Sophia Unsupervised Phonetic Category Learning from Audio and Visual Input |
title | Unsupervised Phonetic Category Learning from
Audio and Visual Input |
title_full | Unsupervised Phonetic Category Learning from
Audio and Visual Input |
title_fullStr | Unsupervised Phonetic Category Learning from
Audio and Visual Input |
title_full_unstemmed | Unsupervised Phonetic Category Learning from
Audio and Visual Input |
title_short | Unsupervised Phonetic Category Learning from
Audio and Visual Input |
title_sort | unsupervised phonetic category learning from audio and visual input |
url | https://hdl.handle.net/1721.1/151659 |
work_keys_str_mv | AT zhisophia unsupervisedphoneticcategorylearningfromaudioandvisualinput |