Speech dataset resources: Difference between revisions
Jump to navigation
Jump to search
No edit summary |
No edit summary |
||
Line 106: | Line 106: | ||
TXT (UTF-8) | TXT (UTF-8) | ||
English conversational speech beyond telephony | |||
|Speakers' gender | |Speakers' gender | ||
Noise/laughter marked | Noise/laughter marked | ||
Line 133: | Line 135: | ||
| | | | ||
| | | | ||
* Cantao Su | |||
* Chenyu Li | |||
* Yanpei Ouyang | |||
* Yi Lei | |||
|- | |- | ||
|[https://paperswithcode.com/dataset/alimeeting AliMeeting (Multi-Channel Multi-Party Meeting Transcription Challenge)] | |[https://paperswithcode.com/dataset/alimeeting AliMeeting (Multi-Channel Multi-Party Meeting Transcription Challenge)] |
Revision as of 11:11, 28 September 2023
This table summarizes dataset resources available
Note: You have editing rights to the table, so you can edit/adjust it to your needs*. Your instructors will clean it up afterwards.
Dataset name | Type of data | Annotation remarks | License | Metadata | Name/s of people who are entering the data | Who are you entering the data? |
---|---|---|---|---|---|---|
LibriSpeech | Audio books (clean, flac format) |
|
Creative Commons Attribution 4.0 International License. |
|
|
|
LibriSpeech | LibriSpeech is a corpus of approximately 1000 hours of read English speech with sampling rate of 16 kHz. The data is derived from read audiobooks from the LibriVox project, and has been carefully segmented and aligned.[1] | Each book’s text is normalized by converting it into upper-case, re- moving the punctuation, and expanding common abbreviations and non-standard words.[2] | CC BY 4.0 |
|
Prepared by Vassil Panayotov with the assistance of Daniel Povey[1] |
The Part 2 of group work is on Librispeech. |
Hugging Face | ||||||
LibriVoxDeEn | Audio based on audio books (.wav file format)
German text and English translation (.tsv file format) |
|
Creative Commons Attribution 4.0 Non-Commercial ShareAlike Internation License |
|
Benjamin Beilharz, Xin Sun, Sariya Karimova, Stefan Riezler | Jocomin Galarneau & Ding Shenghuan & Ömer Tarik Özyilmaz |
SPEECH-COCO | Creative Commons Attribution 4.0 International Liscence |
|
|
|||
SAF (Short Answer Feedback Dataset) | based on audio books (.wav |
|
| |||
ASR-ETELECSC | WAV (PCM)
TXT (UTF-8) English conversational speech beyond telephony |
Speakers' gender
Noise/laughter marked Languages mentioned Starting and ending time of speech Speakers sequenced as numbers [1]/[2] Sound unrecognized as [UNKNOWN] Pause marked as [+] Incomplete words marked as [~] Ambiguity marked as [*] |
MAGIC DATA
OPEN-SOURCE LICENSE |
|
| |
AliMeeting (Multi-Channel Multi-Party Meeting Transcription Challenge) |
|
The annotation is very accurate, but uniquely formeetings like the AliMeeting data, speaker overlap should be explicitly addressed, this question still need to improve. | Creative Commons Attribution ShareAlike 4.0 International License. |
|
Fan Yu, Shiliang Zhang, Pengcheng Guo, Yihui Fu, Zhihao Du, Siqi Zheng, Weilong Huang, Lei Xie, Zheng-Hua Tan, DeLiang Wang, Yanmin Qian, Kong Aik Lee, Zhijie Yan, Bin Ma, Xin Xu, Hui Bu | Dongwen Zhu & Yaling Deng & Chenyi Lin & Soogyeong Shin |
TED-LIUM 3 |
|
Creative Commons BY-NC-ND 3.0. |
|
|
Annie Zhou, XInyi Ma, Jingsi Huang, Igor, Wansu Zhu |
*) this means: you may also add additional pages and link to them in this table
Notes on LibriSpeech
- We only used the development set to test our ASR code in Python;
- The names of the speakers who recorded all the audiobooks contained in this corpus are also available in a separate text file;
- Exhaustive information about this dataset can be found in this article;
- This dataset was included in the Kaldi speech recognition toolkit.
- Another version of research on LibriSpeech by Yuxing Ouyang, Xiaoling Lin, Xueying Liu, Jingxuan Yue, M.Tepei can be found in Librispeech.
References
- ↑ 1.0 1.1 https://www.openslr.org/12
- ↑ Panayotov, V., Chen, G., Povey, D., & Khudanpur, S. (2015, April). Librispeech: an asr corpus based on public domain audio books. In 2015 IEEE international conference on acoustics, speech and signal processing (ICASSP) (pp. 5206-5210). IEEE.
- ↑ Hernandez, François, Vincent Nguyen, Sahar Ghannay, Natalia Tomashenko, and Yannick Estève. “TED-LIUM 3: Twice as Much Data and Corpus Repartition for Experiments on Speaker Adaptation.” In Speech and Computer, edited by Alexey Karpov, Oliver Jokisch, and Rodmonga Potapova, 11096:198–208. Lecture Notes in Computer Science. Cham: Springer International Publishing, 2018. https://doi.org/10.1007/978-3-319-99579-3_21.