mirror of
https://github.com/FunAudioLLM/CosyVoice.git
synced 2026-02-04 17:39:25 +08:00
Merge pull request #62 from passerbya/main
更换默认ttsfrd为WeTextProcessing,修复半角句号结尾或者文本中没有标点会导致合成失败:RuntimeError: torch.cat(): expected a non-empty list of T…
This commit is contained in:
@@ -21,6 +21,8 @@ import torchaudio.compliance.kaldi as kaldi
|
|||||||
import torchaudio
|
import torchaudio
|
||||||
import os
|
import os
|
||||||
import inflect
|
import inflect
|
||||||
|
from tn.chinese.normalizer import Normalizer as ZhNormalizer
|
||||||
|
from tn.english.normalizer import Normalizer as EnNormalizer
|
||||||
try:
|
try:
|
||||||
import ttsfrd
|
import ttsfrd
|
||||||
use_ttsfrd = True
|
use_ttsfrd = True
|
||||||
@@ -61,6 +63,9 @@ class CosyVoiceFrontEnd:
|
|||||||
self.frd.set_lang_type('pinyin')
|
self.frd.set_lang_type('pinyin')
|
||||||
self.frd.enable_pinyin_mix(True)
|
self.frd.enable_pinyin_mix(True)
|
||||||
self.frd.set_breakmodel_index(1)
|
self.frd.set_breakmodel_index(1)
|
||||||
|
else:
|
||||||
|
self.zh_tn_model = ZhNormalizer(remove_erhua=False,full_to_half=False)
|
||||||
|
self.en_tn_model = EnNormalizer()
|
||||||
|
|
||||||
def _extract_text_token(self, text):
|
def _extract_text_token(self, text):
|
||||||
text_token = self.tokenizer.encode(text, allowed_special=self.allowed_special)
|
text_token = self.tokenizer.encode(text, allowed_special=self.allowed_special)
|
||||||
@@ -97,6 +102,8 @@ class CosyVoiceFrontEnd:
|
|||||||
if contains_chinese(text):
|
if contains_chinese(text):
|
||||||
if self.use_ttsfrd:
|
if self.use_ttsfrd:
|
||||||
text = self.frd.get_frd_extra_info(text, 'input')
|
text = self.frd.get_frd_extra_info(text, 'input')
|
||||||
|
else:
|
||||||
|
text = self.zh_tn_model.normalize(text)
|
||||||
text = text.replace("\n", "")
|
text = text.replace("\n", "")
|
||||||
text = replace_blank(text)
|
text = replace_blank(text)
|
||||||
text = replace_corner_mark(text)
|
text = replace_corner_mark(text)
|
||||||
@@ -107,6 +114,7 @@ class CosyVoiceFrontEnd:
|
|||||||
token_min_n=60, merge_len=20,
|
token_min_n=60, merge_len=20,
|
||||||
comma_split=False)]
|
comma_split=False)]
|
||||||
else:
|
else:
|
||||||
|
text = self.en_tn_model.normalize(text)
|
||||||
text = spell_out_number(text, self.inflect_parser)
|
text = spell_out_number(text, self.inflect_parser)
|
||||||
texts = [i for i in split_paragraph(text, partial(self.tokenizer.encode, allowed_special=self.allowed_special), "en", token_max_n=80,
|
texts = [i for i in split_paragraph(text, partial(self.tokenizer.encode, allowed_special=self.allowed_special), "en", token_max_n=80,
|
||||||
token_min_n=60, merge_len=20,
|
token_min_n=60, merge_len=20,
|
||||||
|
|||||||
@@ -74,7 +74,7 @@ def split_paragraph(text: str, tokenize, lang="zh", token_max_n=80, token_min_n=
|
|||||||
return len(tokenize(_text)) < merge_len
|
return len(tokenize(_text)) < merge_len
|
||||||
|
|
||||||
if lang == "zh":
|
if lang == "zh":
|
||||||
pounc = ['。', '?', '!', ';', ':', '.', '?', '!', ';']
|
pounc = ['。', '?', '!', ';', ':', '、', '.', '?', '!', ';']
|
||||||
else:
|
else:
|
||||||
pounc = ['.', '?', '!', ';', ':']
|
pounc = ['.', '?', '!', ';', ':']
|
||||||
if comma_split:
|
if comma_split:
|
||||||
@@ -91,6 +91,11 @@ def split_paragraph(text: str, tokenize, lang="zh", token_max_n=80, token_min_n=
|
|||||||
st = i + 2
|
st = i + 2
|
||||||
else:
|
else:
|
||||||
st = i + 1
|
st = i + 1
|
||||||
|
if len(utts) == 0:
|
||||||
|
if lang == "zh":
|
||||||
|
utts.append(text + '。')
|
||||||
|
else:
|
||||||
|
utts.append(text + '.')
|
||||||
final_utts = []
|
final_utts = []
|
||||||
cur_utt = ""
|
cur_utt = ""
|
||||||
for utt in utts:
|
for utt in utts:
|
||||||
|
|||||||
@@ -25,4 +25,5 @@ soundfile==0.12.1
|
|||||||
tensorboard==2.14.0
|
tensorboard==2.14.0
|
||||||
torch==2.0.1
|
torch==2.0.1
|
||||||
torchaudio==2.0.2
|
torchaudio==2.0.2
|
||||||
wget==3.2
|
wget==3.2
|
||||||
|
WeTextProcessing
|
||||||
|
|||||||
Reference in New Issue
Block a user