FunASR
FunASR copied to clipboard
中文离线转写服务(CPU)如何配置说话人识别模块
Notice: In order to resolve issues more efficiently, please raise issue following the template. (注意:为了更加高效率解决您遇到的问题,请按照模板提问,补充细节)
❓ Questions and Help
中文离线转写服务(CPU)如何配置说话人识别模块, 目前没有看到说话人识别的model配置?
Code
nohup bash run_server.sh
--download-model-dir /workspace/models
--vad-dir damo/speech_fsmn_vad_zh-cn-16k-common-onnx
--model-dir damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-onnx
--punc-dir damo/punc_ct-transformer_zh-cn-common-vocab272727-onnx
--lm-dir damo/speech_ngram_lm_zh-cn-ai-wesp-fst
--itn-dir thuduj12/fst_itn_zh
--certfile 0
--decoder-thread-num 32
--io-thread-num 8
--hotword /workspace/models/hotwords.txt > log.out 2>&1 &
What have you tried?
调用model已经可以实现
from funasr import AutoModel
paraformer-zh is a multi-functional asr model
use vad, punc, spk or not as you need
model = AutoModel(model="paraformer-zh", model_revision="v2.0.4", vad_model="fsmn-vad", vad_model_revision="v2.0.4", punc_model="ct-punc-c", punc_model_revision="v2.0.4", spk_model="cam++", spk_model_revision="v2.0.2", ) res = model.generate(input=f"E:\360MoveData\Users\jll\Desktop\test.wav", batch_size_s=300, hotword='') print(res)
What's your environment?
- OS Windows 11:
- FunASR Version 1.0:
- ModelScope Version (e.g., 1.11.0):
- PyTorch Version (e.g., 2.0.0):
- How you installed funasr pip and source:
- Python version: 3.10
- GPU (e.g., V100M32)
- CUDA/cuDNN version (e.g., cuda11.7):
- Docker version (e.g., funasr-runtime-sdk-cpu-0.4.1)
- Any other relevant information: