Have a personal or library account? Click to login
When Text and Speech are Not Enough: A Multimodal Dataset of Collaboration in a Situated Task Cover

Figures & Tables

Table 1

Participant pool distribution of gender and ethnic background. The task was conducted in English. Native languages besides English included Assamese, Bengali, Gujarati, Hindi, Malayalam, Persian, Spanish, Telugu, and Urdu.

MALEFEMALECAUCASIAN NON-HISPANICHISPANIC/LATINOASIAN
80%20%60%10%30%
Table 2

Dataset descriptive statistics.

AVG.SDMIN.MAX.
Participant age (yrs.)24.584.581935
Video length (mins.)17.007.00934
johd-10-168-g1.jpg
Figure 1

Three participants engaged in the Weights Task. Participant #3 (on the right) is taking a block off the scale to try another configuration while Participant #2 (in the middle) wants to clarify the weight of the block under it. Multimodal information is required to make such a judgment.

johd-10-168-g2.jpg
Figure 2

Multichannel (GAMR, NICE, speech transcription, and CPS) annotation “score” using ELAN (Brugman & Russel, 2004).

DOI: https://doi.org/10.5334/johd.168 | Journal eISSN: 2059-481X
Language: English
Submitted on: Oct 14, 2023
Accepted on: Dec 5, 2023
Published on: Jan 17, 2024
Published by: Ubiquity Press
In partnership with: Paradigm Publishing Services
Publication frequency: 1 issue per year

© 2024 Ibrahim Khebour, Richard Brutti, Indrani Dey, Rachel Dickler, Kelsey Sikes, Kenneth Lai, Mariah Bradford, Brittany Cates, Paige Hansen, Changsoo Jung, Brett Wisniewski, Corbyn Terpstra, Leanne Hirshfield, Sadhana Puntambekar, Nathaniel Blanchard, James Pustejovsky, Nikhil Krishnaswamy, published by Ubiquity Press
This work is licensed under the Creative Commons Attribution 4.0 License.