|
import datasets |
|
import collections |
|
import gzip |
|
import textwrap |
|
|
|
_DESCRIPTION = "tba" |
|
_URL = "tba" |
|
_CITATION = "tba" |
|
_LICENSE = "tba" |
|
|
|
class VoxLinguaConfig(datasets.BuilderConfig): |
|
"""VoxLingua107 corpus.""" |
|
def __init__( |
|
self, |
|
features, |
|
url, |
|
data_url=None, |
|
supervised_keys=None, |
|
shuffled=False, |
|
deduplicated=False, |
|
**kwargs, |
|
): |
|
super(VoxLinguaConfig, self).__init__(version=datasets.Version("1.9.0", ""), **kwargs) |
|
self.features = features |
|
self.data_url = data_url |
|
self.url = url |
|
self.supervised_keys = supervised_keys |
|
|
|
def _languages(): |
|
"""Create the sorted dictionary of language codes, and language names. |
|
Returns: |
|
The sorted dictionary as an instance of `collections.OrderedDict`. |
|
""" |
|
langs = { |
|
"af":"", |
|
"am":"", |
|
"ar":"", |
|
"as":"", |
|
"az":"", |
|
"ba":"", |
|
"be":"", |
|
"bg":"", |
|
"bn":"", |
|
"bo":"", |
|
"br":"", |
|
"bs":"", |
|
"ca":"", |
|
"ceb":"", |
|
"cs":"", |
|
"cy":"", |
|
"da":"", |
|
"de":"", |
|
"el":"", |
|
"en":"", |
|
"eo":"", |
|
"es":"", |
|
"et":"", |
|
"eu":"", |
|
"fa":"", |
|
"fi":"", |
|
"fo":"", |
|
"fr":"", |
|
"gl":"", |
|
"gn":"", |
|
"gu":"", |
|
"gv":"", |
|
"ha":"", |
|
"haw":"", |
|
"hi":"", |
|
"hr":"", |
|
"ht":"", |
|
"hu":"", |
|
"hy":"", |
|
"ia":"", |
|
"id":"", |
|
"is":"", |
|
"it":"", |
|
"iw":"", |
|
"ja":"", |
|
"jw":"", |
|
"ka":"", |
|
"kk":"", |
|
"km":"", |
|
"kn":"", |
|
"ko":"", |
|
"la":"", |
|
"lb":"", |
|
"ln":"", |
|
"lo":"", |
|
"lt":"", |
|
"lv":"", |
|
"mg":"", |
|
"mi":"", |
|
"mk":"", |
|
"ml":"", |
|
"mn":"", |
|
"mr":"", |
|
"ms":"", |
|
"mt":"", |
|
"my":"", |
|
"ne":"", |
|
"nl":"", |
|
"nn":"", |
|
"no":"", |
|
"oc":"", |
|
"pa":"", |
|
"pl":"", |
|
"ps":"", |
|
"pt":"", |
|
"ro":"", |
|
"ru":"", |
|
"sa":"", |
|
"sco":"", |
|
"sd":"", |
|
"si":"", |
|
"sk":"", |
|
"sl":"", |
|
"sn":"", |
|
"so":"", |
|
"sq":"", |
|
"sr":"", |
|
"su":"", |
|
"sv":"", |
|
"sw":"", |
|
"ta":"", |
|
"te":"", |
|
"tg":"", |
|
"th":"", |
|
"tk":"", |
|
"tl":"", |
|
"tr":"", |
|
"tt":"", |
|
"uk":"", |
|
"ur":"", |
|
"uz":"", |
|
"vi":"", |
|
"war":"", |
|
"yi":"", |
|
"yo":"", |
|
"zh":"" |
|
} |
|
return collections.OrderedDict(sorted(langs.items())) |
|
|
|
|
|
class VoxLingua(datasets.GeneratorBasedBuilder): |
|
BUILDER_CONFIGS = [ |
|
VoxLinguaConfig( |
|
name = language, |
|
description=textwrap.dedent( |
|
""" tbd """ |
|
), |
|
shuffled=False, |
|
deduplicated=False, |
|
features=datasets.Features( |
|
{ |
|
"file": datasets.Value("string"), |
|
"audio": datasets.Audio(sampling_rate=16_000), |
|
"label": datasets.ClassLabel( |
|
names=[f"{i}" for i in range(107)] |
|
), |
|
} |
|
), |
|
supervised_keys=("file", "label"), |
|
url="http://bark.phon.ioc.ee/voxlingua107/", |
|
data_url="http://bark.phon.ioc.ee/voxlingua107/{language}.zip" |
|
) |
|
for language in _languages() |
|
] |
|
|
|
BUILDER_CONFIG_CLASS = VoxLinguaConfig |
|
|
|
def _info(self): |
|
return datasets.DatasetInfo( |
|
description=_DESCRIPTION, |
|
features=self.config.features, |
|
supervised_keys=self.config.supervised_keys, |
|
homepage=self.config.url, |
|
citation=_CITATION, |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
train_data_urls = [self.config.url + f"{key}.zip" for key in _languages().keys()] |
|
downloaded_files_train = dl_manager.download(train_data_urls) |
|
dev_data_url = [self.config.url + f"dev.zip"] |
|
downloaded_files_dev = dl_manager.download(dev_data_url) |
|
return [ |
|
datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"archive_path": downloaded_files_train}), |
|
datasets.SplitGenerator(name=datasets.Split.DEV, gen_kwargs={"archive_path": downloaded_files_dev}), |
|
] |
|
|
|
def _generate_examples(self, archive_path, split=None): |
|
return "" |
|
|
|
|