emotion, vocal emotion, acting, psychology, neural networks, artificial intelligence, voice

Data Set Description Summary

This corpus contains over 50 hours of voice acted readings as part of a dissertation project. These recordings represent one of four emotions or the subject's normal speaking voice. The four emotions acted are: anger, fear, happiness, and sadness. These recordings can be useful for building a simple emotion recognition model. Data were collected from BYU students in 2019. Supporting documents designed for people wanting replicate this project are included in For more information about these files, read the README file.

All audio files are encoded as 44.1 kHz .wav mono audio. Each subject read the 50-word script multiple times, with one intact reading found in each recording. represents the most complete and polished set of data, with 120 subjects represented in 11,831 .wav files. are from these same 120 subjects, but in a separate task, which involved reading the script while acting one emotion in as many different ways as possible. This folder contains 997 .wav files.

Various script-reading issues are found in for the same 120 subjects.

An additional 11 subjects are included in, who have too few good recordings in each emotion, prevalent background noise, or had the research assistant end the recording session prematurely.

Additional audio files are found in, which contains 168 files that failed one analysis step. contains information for timing of words and sounds in each audio file.

This data set is licensed under CC BY-NC-SA 4.0, explained in the LICENSE file.

Document Type


Publication Date


Data Collection Start Date


Data Collection End Date





Family, Home, and Social Sciences



University Standing at time of data collection

Graduate Student

Creative Commons License

Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 License.