|
@@ -4,12 +4,13 @@ from modelscope.utils.constant import Tasks
|
|
|
if __name__ == '__main__':
|
|
if __name__ == '__main__':
|
|
|
audio_in = 'https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav'
|
|
audio_in = 'https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav'
|
|
|
output_dir = None
|
|
output_dir = None
|
|
|
- inference_pipline = pipeline(
|
|
|
|
|
|
|
+ inference_pipeline = pipeline(
|
|
|
task=Tasks.auto_speech_recognition,
|
|
task=Tasks.auto_speech_recognition,
|
|
|
- model="damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch",
|
|
|
|
|
- output_dir=output_dir,
|
|
|
|
|
- batch_size=1,
|
|
|
|
|
- )
|
|
|
|
|
- rec_result = inference_pipline(audio_in=audio_in)
|
|
|
|
|
|
|
+ model='damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch',
|
|
|
|
|
+ vad_model='damo/speech_fsmn_vad_zh-cn-16k-common-pytorch',
|
|
|
|
|
+ vad_model_revision="v1.1.8",
|
|
|
|
|
+ punc_model='damo/punc_ct-transformer_zh-cn-common-vocab272727-pytorch',
|
|
|
|
|
+ punc_model_revision="v1.1.6")
|
|
|
|
|
+ rec_result = inference_pipeline(audio_in=audio_in)
|
|
|
print(rec_result)
|
|
print(rec_result)
|
|
|
|
|
|