|
import os |
|
import sys |
|
from tqdm import tqdm |
|
from ipatok import tokenise |
|
from glob import glob |
|
from lhotse import CutSet |
|
from lhotse.shar.writers import SharWriter |
|
from pathlib import Path |
|
import logging |
|
|
|
|
|
logging.basicConfig( |
|
level=logging.DEBUG, |
|
format="[%(asctime)s] %(levelname)s [%(name)s.%(funcName)s:%(lineno)d] %(message)s", |
|
datefmt="%Y/%b/%d %H:%M:%S", |
|
stream=sys.stdout) |
|
|
|
|
|
|
|
inpath = sys.argv[1] |
|
outpath = sys.argv[2] |
|
|
|
filelist = glob(inpath+'/**/*.jsonl.gz',recursive=True) |
|
|
|
|
|
datasets = [file.replace(inpath,'') for file in filelist] |
|
datasets = [file.replace(os.path.basename(file),'') for file in datasets] |
|
datasets = list(set(datasets)) |
|
datasets = [file for file in datasets if 'dev' not in file and 'test' not in file and 'doreco' not in file] |
|
print(datasets) |
|
logging.info("%s speech train data files found!"%len(datasets)) |
|
|
|
logging.info("Beginning processing dataset") |
|
|
|
|
|
data_dir = Path(outpath) |
|
data_dir.mkdir(parents=True, exist_ok=True) |
|
with SharWriter(data_dir, fields={"recording": "flac"}, shard_size=20000) as writer: |
|
|
|
for i,dataset in enumerate(datasets): |
|
|
|
data_path = inpath+dataset |
|
logging.info("Processing %s"%data_path) |
|
|
|
supervision = sorted(glob(os.path.join(data_path,'cuts*'))) |
|
recording = sorted(glob(os.path.join(data_path,'recording*'))) |
|
assert len(supervision)==len(recording) |
|
|
|
logging.info("%s shards found"%len(supervision)) |
|
|
|
cuts = CutSet.from_shar( |
|
{ |
|
"cuts": supervision, |
|
"recording": recording |
|
} |
|
) |
|
|
|
|
|
for cut in tqdm(cuts): |
|
writer.write(cut) |
|
|
|
logging.info("Processing done! %s datasets remaining."%(len(datasets)-i-1)) |
|
|