Source code for lhotse.recipes.xbmu_amdo31

"""
About the XBMU-AMDO31 corpus
XBMU-AMDO31 is an open-source Amdo Tibetan speech corpus published by Northwest Minzu University.
publicly available on https://huggingface.co/datasets/syzym/xbmu_amdo31

XBMU-AMDO31 dataset is a speech recognition corpus of Amdo Tibetan dialect.
The open source corpus contains 31 hours of speech data and resources related
to build speech recognition systems,including transcribed texts and a Tibetan
pronunciation lexicon.
(The lexicon is a Tibetan lexicon of the Lhasa dialect, which has been reused
for the Amdo dialect because of the uniformity of the Tibetan language)
The dataset can be used to train a model for Amdo Tibetan Automatic Speech Recognition (ASR).
It was recorded by 66 native speakers of Amdo Tibetan, and the recorded audio was processed and manually inspected.
The dataset has three splits: train, evaluation (dev) and test.Each speaker had approximately 450 sentences,
with a small number of individuals having fewer than 200 sen.

Subset	Hours	Male	Female	Remarks
Train   25.41   27      27      18539 sentences recorded by 54 speakers
Dev     2.81    2       4       2050 sentences recorded by 6 speakers
Test    2.85    3       3       2041 sentences recorded by 6 speakers

Licensing Information
This dataset is distributed under CC BY-SA 4.0.
"""

import logging
import os
import shutil
import tarfile
from collections import defaultdict
from pathlib import Path
from typing import Dict, Optional, Union

from tqdm.auto import tqdm

from lhotse import fix_manifests, validate_recordings_and_supervisions
from lhotse.audio import Recording, RecordingSet
from lhotse.supervision import SupervisionSegment, SupervisionSet
from lhotse.utils import Pathlike, is_module_available, safe_extract


[docs] def download_xbmu_amdo31( target_dir: Pathlike = ".", ) -> Path: """ Downdload and untar the dataset :param target_dir: Pathlike, the path of the dir to storage the dataset. :return: the path to downloaded and extracted directory with data. """ url = f"https://huggingface.co/datasets/syzym/xbmu_amdo31" target_dir = Path(target_dir) target_dir.mkdir(parents=True, exist_ok=True) corpus_dir = target_dir / "xbmu_amdo31" wav_dir = corpus_dir / "data" / "wav" train_tar_name = "train.tar.gz" dev_tar_name = "dev.tar.gz" test_tar_name = "test.tar.gz" if not corpus_dir.is_file(): if is_module_available("git"): from git import Repo else: raise ImportError( "In order to download the xbmu-amdo31 corpus from huggingface, please install the relevant dependencies: pip install gitpython" ) logging.info("Start downloading the xbmu-amdo31 corpus") try: Repo.clone_from(url, corpus_dir) except Exception as error: print(error) raise logging.info("Done") for tar_name in [train_tar_name, dev_tar_name, test_tar_name]: tar_path = wav_dir / tar_name extracted_dir = wav_dir / tar_name[:-7] completed_detector = extracted_dir / ".completed" if completed_detector.is_file(): logging.info(f"Skipping tar of because {completed_detector} exists.") continue shutil.rmtree(extracted_dir, ignore_errors=True) with tarfile.open(tar_path) as tar: safe_extract(tar, path=wav_dir) completed_detector.touch() return corpus_dir
[docs] def prepare_xbmu_amdo31( corpus_dir: Pathlike, output_dir: Optional[Pathlike] = None ) -> Dict[str, Dict[str, Union[RecordingSet, SupervisionSet]]]: """ Returns the manifests which consist of the Recordings and Supervisions :param corpus_dir: Pathlike, the path of the data dir. :param output_dir: Pathlike, the path where to write the manifests. :return: a Dict whose key is the dataset part, and the value is Dicts with the keys 'recordings' and 'supervisions'. """ corpus_dir = Path(corpus_dir) assert corpus_dir.is_dir(), f"No such directory: {corpus_dir}" if output_dir is not None: output_dir = Path(output_dir) output_dir.mkdir(parents=True, exist_ok=True) transcript_path = corpus_dir / "data/transcript/transcript_clean.txt" transcript_dict = {} with open(transcript_path, "r", encoding="utf-8") as f: for line in f.readlines(): idx_transcript = line.split() content = " ".join(idx_transcript[1:]) transcript_dict[idx_transcript[0]] = content manifests = defaultdict(dict) dataset_parts = ["train", "dev", "test"] for part in tqdm( dataset_parts, desc="Process xbmu_amdo31 audio.", ): logging.info(f"Processing xbmu_amdo31 subset: {part}") # Generate a mapping: utt_id -> (audio_path, audio_info, speaker, text) recordings = [] supervisions = [] wav_path = corpus_dir / "data" / "wav" / f"{part}" count = 0 for audio_path in wav_path.rglob("**/*.wav"): idx = audio_path.stem.split("-")[1] speaker = audio_path.parts[-2] if idx not in transcript_dict: logging.warning(f"No transcript: {idx}") logging.warning(f"{audio_path} has no transcript.") continue text = transcript_dict[idx] if not audio_path.is_file(): logging.warning(f"No such file: {audio_path}") continue recording = Recording.from_file(audio_path) recordings.append(recording) count += 1 segment = SupervisionSegment( id=str(count) + "_" + idx, recording_id=speaker + "-" + idx, start=0.0, duration=recording.duration, channel=0, language="tibetan", speaker=speaker, text=text.strip(), ) supervisions.append(segment) recording_set = RecordingSet.from_recordings(recordings) supervision_set = SupervisionSet.from_segments(supervisions) # Fix manifests and validate recording_set, supervision_set = fix_manifests(recording_set, supervision_set) validate_recordings_and_supervisions(recording_set, supervision_set) if output_dir is not None: supervision_set.to_file( output_dir / f"xbmu_amdo31_supervisions_{part}.jsonl.gz" ) recording_set.to_file( output_dir / f"xbmu_amdo31_recordings_{part}.jsonl.gz" ) manifests[part] = {"recordings": recording_set, "supervisions": supervision_set} return manifests