--- license: other license_name: singapore-open-data-license license_link: https://data.gov.sg/open-data-licence task_categories: - text-generation - text-classification - automatic-speech-recognition - audio-classification language: - en pretty_name: The Reprocessed Singapore National Speech Corpus --- # Dataset Card for Reprocessed National Speech Corpus *NOTE: This is an Reprocessed version KaraKaraWitch from Recursal. The official download can be found [here.](https://www.imda.gov.sg/how-we-can-help/national-speech-corpus)* ## Dataset Details ### Dataset Description Dataset Description: The National Speech Corpus (NSC) is the first large-scale Singapore English corpus, sponsored by the Info-communications and Media Development Authority (IMDA) of Singapore. The objective is to serve as a primary resource of open speech data for automatic speech recognition (ASR) research and other applications related to speech processing. Please note that this is an **reprocessed version** of the original corpus available [here](https://www.imda.gov.sg/how-we-can-help/national-speech-corpus). - Curated by: [IMDA Singapore](https://www.imda.gov.sg/how-we-can-help/national-speech-corpus) - Language: Singaporean English - License: [Singapore Open Data License](https://beta.data.gov.sg/open-data-license) - Reprocessed Version: KaraKaraWitch (recursai.ai) This version contains some differences compared to the original because of the unusual file formats, which were hard to process. We modified the formatting and content to make it easier for use in our dataset and for public Huggingface use. For details on the changes and a list of original files and their modified counterparts, please see the documentation in the `Docs` folder. Strike through content denotes the deleted or reformatted information for enhanced usability. All original data has been preserved and made available in the updated formats. ## Changes - Converted docx documentation into markdown text for ease of sharing. The conversation was done with verbatim in mind. Minimizing as much changes as possible. - Converted XLSX tables into utf-8 csv (The BOM is included due to excel's csv saving). - All files are in split tar files in standard json / jsonl files + FLAC Compressed audio. - We have not modified any text transcripts, audio files were originally saved in .wav which we have losslessly compressed to flac. ## Uses The National Speech Corpus (NSC) is a large-scale Singapore English corpus aimed to improve speech engines' accuracy of recognition and transcription for locally accented English, supporting innovative digital solutions and driving progress in Singapore's digital landscape. ### Direct Use The NSC can be used to improve Automatic Speech Recognition (ASR) research and speech-related applications, such as telco call centres to transcribe calls for auditing and sentiment analysis purposes, and chatbots that can accurately support the Singaporean accent. ### Out-of-Scope Use The NSC is not intended for speech synthesis technology, but it can contribute to producing an AI voice that is more familiar to Singaporeans, with local terms pronounced more accurately. ## Dataset Structure The entire NSC is approximately 1.2 TB (~800GB with FLAC Compression and json metadata) in size, consisting of 6 parts. **Due to HF file restrictions**, we have chunked the file with `split` command. As such you will need to recombine the files with `cat` or similar methods. Using `DifferentRooms.tar` from `Part 3` as an Example: `cat "DifferentRooms-00.tar" "DifferentRooms-01.tar" [...] "DifferentRooms-06.tar" "DifferentRooms-07.tar" > "DifferentRooms.tar"` After concating / combining the files, refer to the `Usage Example` on how to use it. ## Usage Example The dataset can be loaded with webdataset. ```py import webdataset as wds # After concatting, you may use the file like a regular dataset. # The dataset is compatible with WebDataset format. Example... # filepath to Part 1/Channel_0.tar FILEPATH = "Channel_0.tar" hf_dataset = wds.WebDataset(FILEPATH).shuffle(1000).to_tuple("json", "flac") for i in hf_dataset: # print(i) # Prints something like this: # i = { # "__key__": "SP0402-CH00-SE00-RC023", # "__url__": FILEPATH, # "json": b'{"SpeakerID":402,"ChannelID":0,"SessionID":0,"RecordingID":23,"original_text":"I felt happy upon hearing the good news from my parents.","read_text":"I felt happy upon hearing the good news from my parents"}', # "flac": b"", # } break ``` If you need to use HF datasets, load it like so: ```py from datasets import load_dataset # The tar files are split due to HF limits. You will need to combine them first. # You may use the following # `cat "DifferentRooms-00.tar" "DifferentRooms-01.tar" ... "DifferentRooms-06.tar" "DifferentRooms-07.tar" > "DifferentRooms.tar"` # After concatting, you may use the file like a regular dataset. FILEPATH = "Channel_0.tar" hf_dataset = load_dataset("webdataset", data_files={"train": FILEPATH}, split="train", streaming=True) # NOTE: You will need to install 'librosa' and 'soundfile' to decode the flac file. for i in hf_dataset: print(i) # Prints something like this: # { # "__key__": "SP0402-CH00-SE00-RC001", # "__url__": FILEPATH, # "json": { # "ChannelID": 0, # "RecordingID": 1, # "SessionID": 0, # "SpeakerID": 402, # "original_text": "Mary and her family were moving to another city.", # "read_text": "Mary and her family were moving to another city", # }, # "flac": { # "path": "SP0402-CH00-SE00-RC001.flac", # "array": array( # [ # 0.00000000e00, # 6.10351562e-05, # 1.52587891e-04, # ..., # -2.44140625e-04, # -2.44140625e-04, # -1.83105469e-04, # ] # ), # "sampling_rate": 16000, # }, # } break ``` ## Other Notes - Downloading from dropbox is really not optimal. I've sent a email to `nsc@imda.gov.sg` but they didn't respond to me. 😔 - Managed to get a response from them. Though it is around 1 month and I've already downloaded all the files from the dropbox link. - The scripts has numerous text issues. Notably missing quotes and incorrect encoding which made it a absolute headache to process. - I'll probably release the tools at a later time. ## BibTeX Citation ```tex @ONLINE{reprocessed_nationalspeechcorpus, title = {The Reprocessed National Speech Corpus}, author = {KaraKaraWitch}, year = {2024}, howpublished = {\url{https://huggingface.co/datasets/recursal/reprocessed_national_speech_corpus}}, } @ONLINE{imda_nationalspeechcorpus, title = {IMDA National Speech Corpus}, author = {Infocomm Media Development Authority}, year = {2024}, howpublished = {\url{https://www.imda.gov.sg/how-we-can-help/national-speech-corpus}}, } ``` ## Glossary There is a compiled markdown containing documents and notes various notes. ## Recursal's Vision > To make AI accessible to everyone, regardless of language, or economical status This is the collective goal of the `RWKV Open Source foundation` and `Recursal AI`, the commercial entity who backs it. We believe that AI should not be controlled by a select few individual organization. And that it should be made accessible regardless if you are rich or poor, or a native speaker of english. ### About RWKV RWKV is an Open Source, non profit group, under the linux foundation. Focused on developing the RWKV AI architecture, in accordence to our vision. The RWKV architecture scales efficiently and economically. As an RNN & Transformer hybrid, it is able to provide the performance similar to leading transformer models, while having the compute and energy efficiency of an RNN based architecture. You can find out more about the project, and latest models, at the following - [https://blog.rwkv.com](https://blog.rwkv.com) - [https://wiki.rwkv.com](https://wiki.rwkv.com) ### About Recursal AI Recursal AI, is the commercial entity built to provide support for RWKV model development and users, while providing commercial services via its public cloud, or private-cloud / on-premise offerings. As part of our vision. Our commitment, is to ensure open source development and access to the best foundational AI models and datasets. The following dataset/models provided here, is part of that commitment. You can find out more about recursal AI here - [https://recursal.ai](https://recursal.ai) - [https://blog.recursal.ai](https://blog.recursal.ai) ## Dataset Card Contact For issues regarding this reprocessed dataset, you may use the **community discussions** thread. In regards to licensing, please refer to *Singapore Open Data License* listed above. **We will close any issues regard licensing** due to abuse as of late due to it. For further enquiries on the **original** National Speech Corpus (raw data), please contact ****.