Reaching beneath the tip of the iceberg: A guide to the Freiburg Multimodal Interaction Corpus

Most corpora tacitly subscribe to a speech-only view filtering out anything that is not a ‘word’ and transcribing the spoken language merely orthographically despite the fact that the “speech-only view on language is fundamentally incomplete” (Kok 2017, 2) due to the deep intertwining of the verbal,...

Full description

Bibliographic Details
Main Authors: Rühlemann Christoph, Ptak Alexander
Format: Article
Language:English
Published: De Gruyter 2023-11-01
Series:Open Linguistics
Subjects:
Online Access:https://doi.org/10.1515/opli-2022-0245
_version_ 1797561943913922560
author Rühlemann Christoph
Ptak Alexander
author_facet Rühlemann Christoph
Ptak Alexander
author_sort Rühlemann Christoph
collection DOAJ
description Most corpora tacitly subscribe to a speech-only view filtering out anything that is not a ‘word’ and transcribing the spoken language merely orthographically despite the fact that the “speech-only view on language is fundamentally incomplete” (Kok 2017, 2) due to the deep intertwining of the verbal, vocal, and kinesic modalities (Levinson and Holler 2014). This article introduces the Freiburg Multimodal Interaction Corpus (FreMIC), a multimodal and interactional corpus of unscripted conversation in English currently under construction. At the time of writing, FreMIC comprises (i) c. 29 h of video-recordings transcribed and annotated in detail and (ii) automatically (and manually) generated multimodal data. All conversations are transcribed in ELAN both orthographically and using Jeffersonian conventions to render verbal content and interactionally relevant details of sequencing (e.g. overlap, latching), temporal aspects (pauses, acceleration/deceleration), phonological aspects (e.g. intensity, pitch, stretching, truncation, voice quality), and laughter. Moreover, the orthographic transcriptions are exhaustively PoS-tagged using the CLAWS web tagger (Garside and Smith 1997). ELAN-based transcriptions also provide exhaustive annotations of re-enactments (also referred to as (free) direct speech, constructed dialogue, etc.) as well as silent gestures (meaningful gestures that occur without accompanying speech). The multimodal data are derived from psychophysiological measurements and eye tracking. The psychophysiological measurements include, inter alia, electrodermal activity or GSR, which is indicative of emotional arousal (e.g. Peräkylä et al. 2015). Eye tracking produces data of two kinds: gaze direction and pupil size. In FreMIC, gazes are automatically recorded using the area-of-interest technology. Gaze direction is interactionally key, for example, in turn-taking (e.g. Auer 2021) and re-enactments (e.g. Pfeiffer and Weiss 2022), while changes in pupil size provide a window onto cognitive intensity (e.g. Barthel and Sauppe 2019). To demonstrate what opportunities FreMIC’s (combination of) transcriptions, annotations, and multimodal data open up for research in Interactional (Corpus) Linguistics, this article reports on interim results derived from work-in-progress.
first_indexed 2024-03-10T18:21:47Z
format Article
id doaj.art-5e2f7e57797e41fc892aa24ff6545f5f
institution Directory Open Access Journal
issn 2300-9969
language English
last_indexed 2024-03-10T18:21:47Z
publishDate 2023-11-01
publisher De Gruyter
record_format Article
series Open Linguistics
spelling doaj.art-5e2f7e57797e41fc892aa24ff6545f5f2023-11-20T07:16:44ZengDe GruyterOpen Linguistics2300-99692023-11-0191p. 19723110.1515/opli-2022-0245Reaching beneath the tip of the iceberg: A guide to the Freiburg Multimodal Interaction CorpusRühlemann Christoph0Ptak Alexander1Germanistische Linguistik, Freiburg University, Freiburg, GermanyGermanistische Linguistik, Freiburg University, Freiburg, GermanyMost corpora tacitly subscribe to a speech-only view filtering out anything that is not a ‘word’ and transcribing the spoken language merely orthographically despite the fact that the “speech-only view on language is fundamentally incomplete” (Kok 2017, 2) due to the deep intertwining of the verbal, vocal, and kinesic modalities (Levinson and Holler 2014). This article introduces the Freiburg Multimodal Interaction Corpus (FreMIC), a multimodal and interactional corpus of unscripted conversation in English currently under construction. At the time of writing, FreMIC comprises (i) c. 29 h of video-recordings transcribed and annotated in detail and (ii) automatically (and manually) generated multimodal data. All conversations are transcribed in ELAN both orthographically and using Jeffersonian conventions to render verbal content and interactionally relevant details of sequencing (e.g. overlap, latching), temporal aspects (pauses, acceleration/deceleration), phonological aspects (e.g. intensity, pitch, stretching, truncation, voice quality), and laughter. Moreover, the orthographic transcriptions are exhaustively PoS-tagged using the CLAWS web tagger (Garside and Smith 1997). ELAN-based transcriptions also provide exhaustive annotations of re-enactments (also referred to as (free) direct speech, constructed dialogue, etc.) as well as silent gestures (meaningful gestures that occur without accompanying speech). The multimodal data are derived from psychophysiological measurements and eye tracking. The psychophysiological measurements include, inter alia, electrodermal activity or GSR, which is indicative of emotional arousal (e.g. Peräkylä et al. 2015). Eye tracking produces data of two kinds: gaze direction and pupil size. In FreMIC, gazes are automatically recorded using the area-of-interest technology. Gaze direction is interactionally key, for example, in turn-taking (e.g. Auer 2021) and re-enactments (e.g. Pfeiffer and Weiss 2022), while changes in pupil size provide a window onto cognitive intensity (e.g. Barthel and Sauppe 2019). To demonstrate what opportunities FreMIC’s (combination of) transcriptions, annotations, and multimodal data open up for research in Interactional (Corpus) Linguistics, this article reports on interim results derived from work-in-progress.https://doi.org/10.1515/opli-2022-0245talk-in-interactionmultimodalitygestureeye trackingcorpus construction
spellingShingle Rühlemann Christoph
Ptak Alexander
Reaching beneath the tip of the iceberg: A guide to the Freiburg Multimodal Interaction Corpus
Open Linguistics
talk-in-interaction
multimodality
gesture
eye tracking
corpus construction
title Reaching beneath the tip of the iceberg: A guide to the Freiburg Multimodal Interaction Corpus
title_full Reaching beneath the tip of the iceberg: A guide to the Freiburg Multimodal Interaction Corpus
title_fullStr Reaching beneath the tip of the iceberg: A guide to the Freiburg Multimodal Interaction Corpus
title_full_unstemmed Reaching beneath the tip of the iceberg: A guide to the Freiburg Multimodal Interaction Corpus
title_short Reaching beneath the tip of the iceberg: A guide to the Freiburg Multimodal Interaction Corpus
title_sort reaching beneath the tip of the iceberg a guide to the freiburg multimodal interaction corpus
topic talk-in-interaction
multimodality
gesture
eye tracking
corpus construction
url https://doi.org/10.1515/opli-2022-0245
work_keys_str_mv AT ruhlemannchristoph reachingbeneaththetipoftheicebergaguidetothefreiburgmultimodalinteractioncorpus
AT ptakalexander reachingbeneaththetipoftheicebergaguidetothefreiburgmultimodalinteractioncorpus