Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

【TTS】启动英文流式服务报错 #3733

Closed
Ankh-L opened this issue Apr 1, 2024 · 3 comments
Closed

【TTS】启动英文流式服务报错 #3733

Ankh-L opened this issue Apr 1, 2024 · 3 comments
Labels

Comments

@Ankh-L
Copy link

Ankh-L commented Apr 1, 2024

根据中文的demo改了一下am和voc,启动失败,看报错好像只支持中文流式?下面是我的配置:
``
#################################################################################

SERVER SETTING

#################################################################################
host: 0.0.0.0
port: 8888

The task format in the engin_list is: _

engine_list choices = ['tts_online', 'tts_online-onnx'], the inference speed of tts_online-onnx is faster than tts_online.

protocol choices = ['websocket', 'http']

protocol: 'websocket'
engine_list: ['tts_online-onnx']
#################################################################################

ENGINE CONFIG

#################################################################################

################################### TTS #########################################
################### speech task: tts; engine_type: online-onnx #######################
tts_online-onnx:
# am (acoustic model) choices=['fastspeech2_csmsc_onnx', 'fastspeech2_cnndecoder_csmsc_onnx']
# fastspeech2_cnndecoder_csmsc_onnx support streaming am infer.
am: 'fastspeech2_ljspeech_onnx'
# am_ckpt is a list, if am is fastspeech2_cnndecoder_csmsc_onnx, am_ckpt = [encoder model, decoder model, postnet model];
# if am is fastspeech2_csmsc_onnx, am_ckpt = [ckpt model];
am_ckpt: # list
am_stat:
phones_dict:
tones_dict:
speaker_dict:
am_sample_rate: 24000
am_sess_conf:
device: "cpu" # set 'gpu:id' or 'cpu'
use_trt: False
cpu_threads: 4

# voc (vocoder) choices=['mb_melgan_csmsc_onnx, hifigan_csmsc_onnx']
# Both mb_melgan_csmsc_onnx and hifigan_csmsc_onnx support streaming voc inference
voc: 'hifigan_ljspeech_onnx'
voc_ckpt:
voc_sample_rate: 24000
voc_sess_conf:
    device: "cpu" # set 'gpu:id' or 'cpu'
    use_trt: False
    cpu_threads: 4

# others
lang: 'en'
# am_block and am_pad only for fastspeech2_cnndecoder_onnx model to streaming am infer,
# when am_pad set 12, streaming synthetic audio is the same as non-streaming synthetic audio
am_block: 72
am_pad: 10
# voc_pad and voc_block voc model to streaming voc infer,
# when voc model is mb_melgan_csmsc_onnx, voc_pad set 14, streaming synthetic audio is the same as non-streaming synthetic audio; The minimum value of pad can be set to 7, streaming synthetic audio sounds normal
# when voc model is hifigan_csmsc_onnx, voc_pad set 19, streaming synthetic audio is the same as non-streaming synthetic audio; voc_pad set 14, streaming synthetic audio sounds normal
voc_block: 36
voc_pad: 7
# voc_upsample should be same as n_shift on voc config.
voc_upsample: 300

``

报错信息:
[2024-04-01 07:29:24,459] [ ERROR] - Please check config, am support: fastspeech2, voc support: hifigan_csmsc-zh or mb_melgan_csmsc.

@Ankh-L Ankh-L added the Question label Apr 1, 2024
@Ray961123
Copy link

开发者你好,感谢关注 PaddleSpeech 开源项目,抱歉给你带来了不好的开发体验,目前开源项目维护人力有限,你可以尝试通过修改 PaddleSpeech 源码的方式自己解决,或请求开源社区其他开发者的协助。飞桨开源社区交流频道:飞桨AI Studio星河社区-人工智能学习与实训社区

@jianghuakun
Copy link

请问解决没?我改成混合模型报错和你一样。

@Ankh-L
Copy link
Author

Ankh-L commented May 24, 2024

不支持英文流式 @jianghuakun

@Ankh-L Ankh-L closed this as completed May 24, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

3 participants