from __future__ import annotations import re from .g2p import * import json import os class Tokenizer(): def __init__(self, data_path: str) -> None: self.dic = {} for line in open(os.path.join(data_path, "dictionary.txt")): items = line.split() self.dic[items[0]] = " ".join(items[1:]) self.config = json.load(open(os.path.join(data_path, "config.json"))) def g2p(self, text): text = re.sub("—", "-", text) text = re.sub("([!'(),-.:;?])", r' \1 ', text) phonemes = [] for word in text.split(): if re.match("[!'(),-.:;?]", word): phonemes.append(word) continue word = word.lower() if len(phonemes) > 0: phonemes.append(' ') if word in self.dic: phonemes.extend(self.dic[word].split()) else: phonemes.extend(convert(word).split()) phoneme_id_map = self.config["phoneme_id_map"] phoneme_ids = [] phoneme_ids.extend(phoneme_id_map["^"]) phoneme_ids.extend(phoneme_id_map["_"]) for p in phonemes: if p in phoneme_id_map: phoneme_ids.extend(phoneme_id_map[p]) phoneme_ids.extend(phoneme_id_map["_"]) phoneme_ids.extend(phoneme_id_map["$"]) return phoneme_ids, phonemes def _get_seq(self, text: str) -> list[int]: seq = self.g2p(text)[0] return seq