From c93d3dda01ae69fde8a4b2372f0e4260135599c3 Mon Sep 17 00:00:00 2001 From: "lyuxiang.lx" Date: Thu, 29 Jan 2026 17:31:04 +0000 Subject: [PATCH] update --- cosyvoice/dataset/processor.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/cosyvoice/dataset/processor.py b/cosyvoice/dataset/processor.py index 9893410..deba209 100644 --- a/cosyvoice/dataset/processor.py +++ b/cosyvoice/dataset/processor.py @@ -405,7 +405,7 @@ def padding(data, use_spk_embedding, mode='train', gan=False, dpo=False): batch['instruct_token_len'] = torch.tensor([i.size(0) for i in instruct_token], dtype=torch.int32) batch['instruct_token'] = pad_sequence(instruct_token, batch_first=True, padding_value=0) if torch.tensor(['whisper_feat' in sample[i] for i in order]).all(): - whisper_feat = [torch.tensor(sample[i]['whisper_feat']) for i in order] + whisper_feat = [sample[i]['whisper_feat'] for i in order] batch['whisper_feat_len'] = torch.tensor([i.size(0) for i in whisper_feat], dtype=torch.int32) batch['whisper_feat'] = pad_sequence(whisper_feat, batch_first=True, padding_value=0) if torch.tensor(['speech_token' in sample[i] for i in order]).all():