test_asr_inference_pipeline.py 23 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436
  1. import unittest
  2. from modelscope.pipelines import pipeline
  3. from modelscope.utils.constant import Tasks
  4. from modelscope.utils.logger import get_logger
  5. logger = get_logger()
  6. class TestConformerInferencePipelines(unittest.TestCase):
  7. def test_funasr_path(self):
  8. import funasr
  9. import os
  10. logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
  11. def test_aishell1(self):
  12. inference_pipeline = pipeline(
  13. task=Tasks.auto_speech_recognition,
  14. model='damo/speech_conformer_asr_nat-zh-cn-16k-aishell1-vocab4234-pytorch')
  15. rec_result = inference_pipeline(
  16. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  17. logger.info("asr inference result: {0}".format(rec_result))
  18. def test_aishell2(self):
  19. inference_pipeline = pipeline(
  20. task=Tasks.auto_speech_recognition,
  21. model='damo/speech_conformer_asr_nat-zh-cn-16k-aishell2-vocab5212-pytorch')
  22. rec_result = inference_pipeline(
  23. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  24. logger.info("asr inference result: {0}".format(rec_result))
  25. class TestData2vecInferencePipelines(unittest.TestCase):
  26. def test_funasr_path(self):
  27. import funasr
  28. import os
  29. logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
  30. def test_transformer(self):
  31. inference_pipeline = pipeline(
  32. task=Tasks.auto_speech_recognition,
  33. model='damo/speech_data2vec_pretrain-zh-cn-aishell2-16k-pytorch')
  34. rec_result = inference_pipeline(
  35. audio_in='https://modelscope.oss-cn-beijing.aliyuncs.com/test/audios/asr_example.wav')
  36. logger.info("asr inference result: {0}".format(rec_result))
  37. def test_paraformer(self):
  38. inference_pipeline = pipeline(
  39. task=Tasks.auto_speech_recognition,
  40. model='damo/speech_data2vec_pretrain-paraformer-zh-cn-aishell2-16k')
  41. rec_result = inference_pipeline(
  42. audio_in='https://modelscope.oss-cn-beijing.aliyuncs.com/test/audios/asr_example.wav')
  43. logger.info("asr inference result: {0}".format(rec_result))
  44. class TestMfccaInferencePipelines(unittest.TestCase):
  45. def test_funasr_path(self):
  46. import funasr
  47. import os
  48. logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
  49. def test_alimeeting(self):
  50. inference_pipeline = pipeline(
  51. task=Tasks.auto_speech_recognition,
  52. model='NPU-ASLP/speech_mfcca_asr-zh-cn-16k-alimeeting-vocab4950',
  53. model_revision='v3.0.0')
  54. rec_result = inference_pipeline(
  55. audio_in='16:32https://pre.modelscope.cn/api/v1/models/NPU-ASLP/speech_mfcca_asr-zh-cn-16k-alimeeting-vocab4950/repo?Revision=master&FilePath=example/asr_example_mc.wav')
  56. logger.info("asr inference result: {0}".format(rec_result))
  57. class TestParaformerInferencePipelines(unittest.TestCase):
  58. def test_funasr_path(self):
  59. import funasr
  60. import os
  61. logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
  62. def test_paraformer_large_contextual_common(self):
  63. param_dict = dict()
  64. param_dict['hotword'] = "https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/hotword.txt"
  65. inference_pipeline = pipeline(
  66. task=Tasks.auto_speech_recognition,
  67. model='damo/speech_paraformer-large-contextual_asr_nat-zh-cn-16k-common-vocab8404',
  68. param_dict=param_dict)
  69. rec_result = inference_pipeline(
  70. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_hotword.wav')
  71. logger.info("asr inference result: {0}".format(rec_result))
  72. def test_paraformer_large_aishell1(self):
  73. inference_pipeline = pipeline(
  74. task=Tasks.auto_speech_recognition,
  75. model='damo/speech_paraformer-large_asr_nat-zh-cn-16k-aishell1-vocab8404-pytorch')
  76. rec_result = inference_pipeline(
  77. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  78. logger.info("asr inference result: {0}".format(rec_result))
  79. def test_paraformer_large_aishell2(self):
  80. inference_pipeline = pipeline(
  81. task=Tasks.auto_speech_recognition,
  82. model='damo/speech_paraformer-large_asr_nat-zh-cn-16k-aishell2-vocab8404-pytorch')
  83. rec_result = inference_pipeline(
  84. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  85. logger.info("asr inference result: {0}".format(rec_result))
  86. def test_paraformer_large_common(self):
  87. inference_pipeline = pipeline(
  88. task=Tasks.auto_speech_recognition,
  89. model='damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch')
  90. rec_result = inference_pipeline(
  91. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  92. logger.info("asr inference result: {0}".format(rec_result))
  93. def test_paraformer_tiny_commandword(self):
  94. inference_pipeline = pipeline(
  95. task=Tasks.auto_speech_recognition,
  96. model='damo/speech_paraformer-tiny-commandword_asr_nat-zh-cn-16k-vocab544-pytorch')
  97. rec_result = inference_pipeline(
  98. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh_command.wav')
  99. logger.info("asr inference result: {0}".format(rec_result))
  100. def test_paraformer_8k(self):
  101. inference_pipeline = pipeline(
  102. task=Tasks.auto_speech_recognition,
  103. model='damo/speech_paraformer_asr_nat-zh-cn-8k-common-vocab8358-tensorflow1')
  104. rec_result = inference_pipeline(
  105. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_8K.wav')
  106. logger.info("asr inference result: {0}".format(rec_result))
  107. def test_paraformer_aishell1(self):
  108. inference_pipeline = pipeline(
  109. task=Tasks.auto_speech_recognition,
  110. model='damo/speech_paraformer_asr_nat-zh-cn-16k-aishell1-vocab4234-pytorch')
  111. rec_result = inference_pipeline(
  112. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  113. logger.info("asr inference result: {0}".format(rec_result))
  114. def test_paraformer_aishell2(self):
  115. inference_pipeline = pipeline(
  116. task=Tasks.auto_speech_recognition,
  117. model='damo/speech_paraformer_asr_nat-zh-cn-16k-aishell2-vocab5212-pytorch')
  118. rec_result = inference_pipeline(
  119. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  120. logger.info("asr inference result: {0}".format(rec_result))
  121. class TestParaformerBertInferencePipelines(unittest.TestCase):
  122. def test_funasr_path(self):
  123. import funasr
  124. import os
  125. logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
  126. def test_aishell1(self):
  127. inference_pipeline = pipeline(
  128. task=Tasks.auto_speech_recognition,
  129. model='damo/speech_paraformerbert_asr_nat-zh-cn-16k-aishell1-vocab4234-pytorch')
  130. rec_result = inference_pipeline(
  131. audio_in='https://modelscope.oss-cn-beijing.aliyuncs.com/test/audios/asr_example.wav')
  132. logger.info("asr inference result: {0}".format(rec_result))
  133. def test_aishell2(self):
  134. inference_pipeline = pipeline(
  135. task=Tasks.auto_speech_recognition,
  136. model='damo/speech_paraformerbert_asr_nat-zh-cn-16k-aishell2-vocab5212-pytorch')
  137. rec_result = inference_pipeline(
  138. audio_in='https://modelscope.oss-cn-beijing.aliyuncs.com/test/audios/asr_example.wav')
  139. logger.info("asr inference result: {0}".format(rec_result))
  140. class TestUniasrInferencePipelines(unittest.TestCase):
  141. def test_funasr_path(self):
  142. import funasr
  143. import os
  144. logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
  145. def test_uniasr_2pass_cantonese_chs_16k_common_offline(self):
  146. inference_pipeline = pipeline(
  147. task=Tasks.auto_speech_recognition,
  148. model='damo/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-online')
  149. rec_result = inference_pipeline(
  150. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_cantonese-CHS.wav',
  151. param_dict={"decoding_model": "offline"})
  152. logger.info("asr inference result: {0}".format(rec_result))
  153. def test_uniasr_2pass_cantonese_chs_16k_common_online(self):
  154. inference_pipeline = pipeline(
  155. task=Tasks.auto_speech_recognition,
  156. model='damo/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-online')
  157. rec_result = inference_pipeline(
  158. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_cantonese-CHS.wav',
  159. param_dict={"decoding_model": "normal"})
  160. logger.info("asr inference result: {0}".format(rec_result))
  161. def test_uniasr_2pass_cn_dialect_offline(self):
  162. inference_pipeline = pipeline(
  163. task=Tasks.auto_speech_recognition,
  164. model='damo/speech_UniASR_asr_2pass-cn-dialect-16k-vocab8358-tensorflow1-offline')
  165. rec_result = inference_pipeline(
  166. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  167. logger.info("asr inference result: {0}".format(rec_result))
  168. def test_uniasr_2pass_cn_dialect_online(self):
  169. inference_pipeline = pipeline(
  170. task=Tasks.auto_speech_recognition,
  171. model='damo/speech_UniASR_asr_2pass-cn-dialect-16k-vocab8358-tensorflow1-online')
  172. rec_result = inference_pipeline(
  173. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  174. logger.info("asr inference result: {0}".format(rec_result))
  175. def test_uniasr_2pass_de_common_offline(self):
  176. inference_pipeline = pipeline(
  177. task=Tasks.auto_speech_recognition,
  178. model='damo/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-offline')
  179. rec_result = inference_pipeline(
  180. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_de.wav',
  181. param_dict={"decoding_model": "offline"})
  182. logger.info("asr inference result: {0}".format(rec_result))
  183. def test_uniasr_2pass_de_common_online(self):
  184. inference_pipeline = pipeline(
  185. task=Tasks.auto_speech_recognition,
  186. model='damo/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-online')
  187. rec_result = inference_pipeline(
  188. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_de.wav',
  189. param_dict={"decoding_model": "normal"})
  190. logger.info("asr inference result: {0}".format(rec_result))
  191. def test_uniasr_2pass_en_common_offline(self):
  192. inference_pipeline = pipeline(
  193. task=Tasks.auto_speech_recognition,
  194. model='damo/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-offline')
  195. rec_result = inference_pipeline(
  196. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_en.wav',
  197. param_dict={"decoding_model": "offline"})
  198. logger.info("asr inference result: {0}".format(rec_result))
  199. def test_uniasr_2pass_en_common_online(self):
  200. inference_pipeline = pipeline(
  201. task=Tasks.auto_speech_recognition,
  202. model='damo/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-online')
  203. rec_result = inference_pipeline(
  204. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_en.wav',
  205. param_dict={"decoding_model": "normal"})
  206. logger.info("asr inference result: {0}".format(rec_result))
  207. def test_uniasr_2pass_es_common_offline(self):
  208. inference_pipeline = pipeline(
  209. task=Tasks.auto_speech_recognition,
  210. model='damo/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-offline')
  211. rec_result = inference_pipeline(
  212. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_es.wav',
  213. param_dict={"decoding_model": "offline"})
  214. logger.info("asr inference result: {0}".format(rec_result))
  215. def test_uniasr_2pass_es_common_online(self):
  216. inference_pipeline = pipeline(
  217. task=Tasks.auto_speech_recognition,
  218. model='damo/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-online')
  219. rec_result = inference_pipeline(
  220. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_es.wav',
  221. param_dict={"decoding_model": "normal"})
  222. logger.info("asr inference result: {0}".format(rec_result))
  223. def test_uniasr_2pass_fa_common_offline(self):
  224. inference_pipeline = pipeline(
  225. task=Tasks.auto_speech_recognition,
  226. model='damo/speech_UniASR_asr_2pass-fa-16k-common-vocab1257-pytorch-offline')
  227. rec_result = inference_pipeline(
  228. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_fa.wav',
  229. param_dict={"decoding_model": "offline"})
  230. logger.info("asr inference result: {0}".format(rec_result))
  231. def test_uniasr_2pass_fa_common_online(self):
  232. inference_pipeline = pipeline(
  233. task=Tasks.auto_speech_recognition,
  234. model='damo/speech_UniASR_asr_2pass-fa-16k-common-vocab1257-pytorch-online')
  235. rec_result = inference_pipeline(
  236. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_fa.wav',
  237. param_dict={"decoding_model": "normal"})
  238. logger.info("asr inference result: {0}".format(rec_result))
  239. def test_uniasr_2pass_fr_common_offline(self):
  240. inference_pipeline = pipeline(
  241. task=Tasks.auto_speech_recognition,
  242. model='damo/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-offline')
  243. rec_result = inference_pipeline(
  244. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_fr.wav',
  245. param_dict={"decoding_model": "offline"})
  246. logger.info("asr inference result: {0}".format(rec_result))
  247. def test_uniasr_2pass_fr_common_online(self):
  248. inference_pipeline = pipeline(
  249. task=Tasks.auto_speech_recognition,
  250. model='damo/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-online')
  251. rec_result = inference_pipeline(
  252. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_fr.wav',
  253. param_dict={"decoding_model": "normal"})
  254. logger.info("asr inference result: {0}".format(rec_result))
  255. def test_uniasr_2pass_id_common_offline(self):
  256. inference_pipeline = pipeline(
  257. task=Tasks.auto_speech_recognition,
  258. model='damo/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-online')
  259. rec_result = inference_pipeline(
  260. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_id.wav',
  261. param_dict={"decoding_model": "offline"})
  262. logger.info("asr inference result: {0}".format(rec_result))
  263. def test_uniasr_2pass_id_common_online(self):
  264. inference_pipeline = pipeline(
  265. task=Tasks.auto_speech_recognition,
  266. model='damo/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-online')
  267. rec_result = inference_pipeline(
  268. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_id.wav',
  269. param_dict={"decoding_model": "normal"})
  270. logger.info("asr inference result: {0}".format(rec_result))
  271. def test_uniasr_2pass_ja_common_offline(self):
  272. inference_pipeline = pipeline(
  273. task=Tasks.auto_speech_recognition,
  274. model='damo/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-offline')
  275. rec_result = inference_pipeline(
  276. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_ja.wav',
  277. param_dict={"decoding_model": "offline"})
  278. logger.info("asr inference result: {0}".format(rec_result))
  279. def test_uniasr_2pass_ja_common_online(self):
  280. inference_pipeline = pipeline(
  281. task=Tasks.auto_speech_recognition,
  282. model='damo/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-online')
  283. rec_result = inference_pipeline(
  284. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_ja.wav',
  285. param_dict={"decoding_model": "normal"})
  286. logger.info("asr inference result: {0}".format(rec_result))
  287. def test_uniasr_2pass_ko_common_offline(self):
  288. inference_pipeline = pipeline(
  289. task=Tasks.auto_speech_recognition,
  290. model='damo/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-offline')
  291. rec_result = inference_pipeline(
  292. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_ko.wav',
  293. param_dict={"decoding_model": "offline"})
  294. logger.info("asr inference result: {0}".format(rec_result))
  295. def test_uniasr_2pass_ko_common_online(self):
  296. inference_pipeline = pipeline(
  297. task=Tasks.auto_speech_recognition,
  298. model='damo/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-online')
  299. rec_result = inference_pipeline(
  300. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_ko.wav',
  301. param_dict={"decoding_model": "normal"})
  302. logger.info("asr inference result: {0}".format(rec_result))
  303. def test_uniasr_2pass_minnan_common_offline(self):
  304. inference_pipeline = pipeline(
  305. task=Tasks.auto_speech_recognition,
  306. model='damo/speech_UniASR_asr_2pass-minnan-16k-common-vocab3825')
  307. rec_result = inference_pipeline(
  308. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav',
  309. param_dict={"decoding_model": "offline"})
  310. logger.info("asr inference result: {0}".format(rec_result))
  311. def test_uniasr_2pass_pt_common_offline(self):
  312. inference_pipeline = pipeline(
  313. task=Tasks.auto_speech_recognition,
  314. model='damo/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-offline')
  315. rec_result = inference_pipeline(
  316. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_pt.wav',
  317. param_dict={"decoding_model": "offline"})
  318. logger.info("asr inference result: {0}".format(rec_result))
  319. def test_uniasr_2pass_pt_common_online(self):
  320. inference_pipeline = pipeline(
  321. task=Tasks.auto_speech_recognition,
  322. model='damo/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-online')
  323. rec_result = inference_pipeline(
  324. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_pt.wav',
  325. param_dict={"decoding_model": "normal"})
  326. logger.info("asr inference result: {0}".format(rec_result))
  327. def test_uniasr_2pass_ru_common_offline(self):
  328. inference_pipeline = pipeline(
  329. task=Tasks.auto_speech_recognition,
  330. model='damo/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-offline')
  331. rec_result = inference_pipeline(
  332. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_ru.wav',
  333. param_dict={"decoding_model": "offline"})
  334. logger.info("asr inference result: {0}".format(rec_result))
  335. def test_uniasr_2pass_ru_common_online(self):
  336. inference_pipeline = pipeline(
  337. task=Tasks.auto_speech_recognition,
  338. model='damo/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-online')
  339. rec_result = inference_pipeline(
  340. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_ru.wav',
  341. param_dict={"decoding_model": "normal"})
  342. logger.info("asr inference result: {0}".format(rec_result))
  343. def test_uniasr_2pass_vi_common_offline(self):
  344. inference_pipeline = pipeline(
  345. task=Tasks.auto_speech_recognition,
  346. model='damo/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-offline')
  347. rec_result = inference_pipeline(
  348. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_vi.wav',
  349. param_dict={"decoding_model": "offline"})
  350. logger.info("asr inference result: {0}".format(rec_result))
  351. def test_uniasr_2pass_vi_common_online(self):
  352. inference_pipeline = pipeline(
  353. task=Tasks.auto_speech_recognition,
  354. model='damo/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-online')
  355. rec_result = inference_pipeline(
  356. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_vi.wav',
  357. param_dict={"decoding_model": "normal"})
  358. logger.info("asr inference result: {0}".format(rec_result))
  359. def test_uniasr_2pass_zhcn_8k_common_vocab3445_offline(self):
  360. inference_pipeline = pipeline(
  361. task=Tasks.auto_speech_recognition,
  362. model='damo/speech_UniASR_asr_2pass-zh-cn-8k-common-vocab3445-pytorch-offline')
  363. rec_result = inference_pipeline(
  364. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav',
  365. param_dict={"decoding_model": "offline"})
  366. logger.info("asr inference result: {0}".format(rec_result))
  367. def test_uniasr_2pass_zhcn_8k_common_vocab3445_online(self):
  368. inference_pipeline = pipeline(
  369. task=Tasks.auto_speech_recognition,
  370. model='damo/speech_UniASR_asr_2pass-zh-cn-8k-common-vocab3445-pytorch-online')
  371. rec_result = inference_pipeline(
  372. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav',
  373. param_dict={"decoding_model": "normal"})
  374. logger.info("asr inference result: {0}".format(rec_result))
  375. if __name__ == '__main__':
  376. unittest.main()