Parent-Child Vocal Interaction Data
Enter your email below for pricing and additional information. For a faster reply, please contact - contact@deeplyinc.com
Summary
The interactions between 24 pairs of parent and child (48 speakers total), such as reading fairy tales, singing children’s songs, conversing, and others, are contained within this dataset. The recordings took place in 3 different types of places, which include an anechoic chamber, studio apartment, and dance studio, of which the level of reverberation differs. In order to examine the effect of the distance from the mic to the speaker, each interaction is recorded at 3 distinct distances with 2 types of smartphone, iPhone X, and Galaxy S7.
There was one parent and his/her child in a group, and each group is distinguished by a unique key(subject ID). Two speakers(speaker a(parent), speaker b(child)) were sitting on the floor(purple area in the pictures below), and asked to do 3 types of interactions: singing children's song, reading fairy tales, and singing lullabies.
Recording environment : Studio Apartment (moderate reverb), Dance studio (high reverb),
Anechoic Chamber (no reverb)
Device : iPhone X(iOS), Samsung Galaxy S7
Recording distance from the source : 0.4m, 2.0m, 4.0m
Volume(Sample) : ~ 282(~ 3) hours, ~ 360,000(~ 4,000) utterances, ~ 110(~ 0.4) GB
Format : wav/h5(16/44.1kHz, 16-bit, mono)
Language : Korean
Studio Apartment
Dance Studio
Anechoic Chamber
Refer to the dataset descriptions in 'docs' for details and statistics of the entirety of the dataset.
The dataset is a subset (approximately 1%) of a much bigger dataset which were recorded under the same circumstances as these open source samples.
Please contact us (contact@deeplyinc.com) for the pricing and licensing.
Featured Parent Child Vocal Interaction Sound Sample
-
Child reading
-
Child refusing
-
Child singing
-
Parent reading
-
Parent singing
And more sounds!
Dataset statistics
The illustrations below are the statistics about the Deeply Parent-Child Vocal Interaction Dataset. The first two graphs are from the sample dataset, and the others are from the full dataset. To gain more insight about the dataset, please refer to the detailed description in 'docs' and 'Korea_Read_Speech_Corpus.json' in 'Dataset'.
The sample is a partial set of recordings from a single subject group(sub3004), which consists of a 39-year-old female (parent, speaker a) and 5-year-old male (child, speaker b).
Structure
├── dataset
│ ├── AirbnbStudio
│ │ ├── sub30040a00000.wav
│ | └── ...
│ ├── AnechoicChamber
│ │ ├── sub30042a00000.wav
│ | └── ...
│ └── DanceStudio
│ ├── sub30041a00000.wav
│ └── ...
└── docs
├── Deeply\ Parent-Child\ Vocal\ Interaction\ Dataset_Eng.pdf
└── Deeply\ Parent-Child\ Vocal\ Interaction\ Dataset_Kor.pdf
Parent_Child_Vocal_Interaction.json
{'AirbnbStudio':
{'sub30040a00000': {'label': 2,
'subjectID': 'sub3004',
'speaker': 'a',
'age': 39,
'sex': 0,
'noise': 0,
'location': 0,
'distance': 0,
'device': 0,
'rms': 0.005859313067048788,
'length': 1.521},
...
},
...
}
How to decode
label: {speaker a(parent): {0: singing, 1: reading, 2: other utterances},
speaker b(child): {0: singing, 1: reading, 2: crying, 3: refusing, 4: other utterances}}
Subject ID: Unique 'sub + 4-digit' key allocated to each subject group
Speaker: unique key allocated to each individual in the subject group.
Sex: {0: Female, 1: Male}
Noise: {0: Noiseless, 1: Indoor noise, 2: Outdoor noise, 3: Both indoor/outdoor noise}
Location: {0: Studio apartment, 1: Dance studio, 2: Anechoic chamber}
Distance: {0: 0.4m, 1: 2.0m, 2: 4.0m}
Device: {0: iPhone X, 1: Galaxy S7}
Rms: Root mean square value of the signal
Length: length of the signal in seconds
* In polyphonic utterances, the categories, such as label, speaker, and sex, are longer than usual, because we've written the information of both speaker a and b in the same category.
For example, if speaker a(parent, male, 35 yo) sings and speaker b(child, female, 3 yo) tries to talk in a single audio file, speaker would be 'ab', sex would be '10'(male(speaker a), female(speaker b)), and label would be '04'(singing(speaker a), other utterances(speaker b)) . *
License
Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0)
Enter your email below for pricing and additional information. For a faster reply, please contact - contact@deeplyinc.com
Office : E02, Space Sallim 2F, 10, Noryangjin-ro, Dongjak-gu, Seoul, Republic of Korea
Tel : +82 70-7459-0704
E-mail : contact@deeplyinc.com
Copyright © Deeply, Inc. All rights reserved.