|
|
@@ -6,6 +6,28 @@ from modelscope.utils.logger import get_logger
|
|
|
|
|
|
logger = get_logger()
|
|
|
|
|
|
+class TestConformerInferencePipelines(unittest.TestCase):
|
|
|
+ def test_funasr_path(self):
|
|
|
+ import funasr
|
|
|
+ import os
|
|
|
+ logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
|
|
|
+
|
|
|
+ def test_aishell1(self):
|
|
|
+ inference_pipeline = pipeline(
|
|
|
+ task=Tasks.auto_speech_recognition,
|
|
|
+ model='damo/speech_conformer_asr_nat-zh-cn-16k-aishell1-vocab4234-pytorch')
|
|
|
+ rec_result = inference_pipeline(
|
|
|
+ audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
|
|
|
+ logger.info("asr inference result: {0}".format(rec_result))
|
|
|
+
|
|
|
+ def test_aishell2(self):
|
|
|
+ inference_pipeline = pipeline(
|
|
|
+ task=Tasks.auto_speech_recognition,
|
|
|
+ model='damo/speech_conformer_asr_nat-zh-cn-16k-aishell2-vocab5212-pytorch')
|
|
|
+ rec_result = inference_pipeline(
|
|
|
+ audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
|
|
|
+ logger.info("asr inference result: {0}".format(rec_result))
|
|
|
+
|
|
|
class TestData2vecInferencePipelines(unittest.TestCase):
|
|
|
def test_funasr_path(self):
|
|
|
import funasr
|