import os import sys from tqdm import tqdm from ipatok import tokenise from glob import glob from lhotse import CutSet from lhotse.shar.writers import SharWriter from pathlib import Path import logging logging.basicConfig( level=logging.DEBUG, format="[%(asctime)s] %(levelname)s [%(name)s.%(funcName)s:%(lineno)d] %(message)s", datefmt="%Y/%b/%d %H:%M:%S", stream=sys.stdout) inpath = sys.argv[1] outpath = sys.argv[2] filelist = glob(inpath+'/**/*.jsonl.gz',recursive=True) datasets = [file.replace(inpath,'') for file in filelist] datasets = [file.replace(os.path.basename(file),'') for file in datasets] datasets = list(set(datasets)) datasets = [file for file in datasets if 'dev' not in file and 'test' not in file and 'doreco' not in file] print(datasets) logging.info("%s speech train data files found!"%len(datasets)) logging.info("Beginning processing dataset") data_dir = Path(outpath) data_dir.mkdir(parents=True, exist_ok=True) with SharWriter(data_dir, fields={"recording": "flac"}, shard_size=20000) as writer: for i,dataset in enumerate(datasets): data_path = inpath+dataset logging.info("Processing %s"%data_path) supervision = sorted(glob(os.path.join(data_path,'cuts*'))) recording = sorted(glob(os.path.join(data_path,'recording*'))) assert len(supervision)==len(recording) logging.info("%s shards found"%len(supervision)) cuts = CutSet.from_shar( { "cuts": supervision, "recording": recording } ) for cut in tqdm(cuts): writer.write(cut) logging.info("Processing done! %s datasets remaining."%(len(datasets)-i-1))