test_asr_inference_pipeline.py 26 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491
  1. import unittest
  2. from modelscope.pipelines import pipeline
  3. from modelscope.utils.constant import Tasks
  4. from modelscope.utils.logger import get_logger
  5. logger = get_logger()
  6. class TestConformerInferencePipelines(unittest.TestCase):
  7. def test_funasr_path(self):
  8. import funasr
  9. import os
  10. logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
  11. def test_aishell1(self):
  12. inference_pipeline = pipeline(
  13. task=Tasks.auto_speech_recognition,
  14. model='damo/speech_conformer_asr_nat-zh-cn-16k-aishell1-vocab4234-pytorch')
  15. rec_result = inference_pipeline(
  16. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  17. logger.info("asr inference result: {0}".format(rec_result))
  18. def test_aishell2(self):
  19. inference_pipeline = pipeline(
  20. task=Tasks.auto_speech_recognition,
  21. model='damo/speech_conformer_asr_nat-zh-cn-16k-aishell2-vocab5212-pytorch')
  22. rec_result = inference_pipeline(
  23. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  24. logger.info("asr inference result: {0}".format(rec_result))
  25. class TestData2vecInferencePipelines(unittest.TestCase):
  26. def test_funasr_path(self):
  27. import funasr
  28. import os
  29. logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
  30. def test_transformer(self):
  31. inference_pipeline = pipeline(
  32. task=Tasks.auto_speech_recognition,
  33. model='damo/speech_data2vec_pretrain-zh-cn-aishell2-16k-pytorch')
  34. rec_result = inference_pipeline(
  35. audio_in='https://modelscope.oss-cn-beijing.aliyuncs.com/test/audios/asr_example.wav')
  36. logger.info("asr inference result: {0}".format(rec_result))
  37. assert rec_result["text"] == "每一天都要快乐喔"
  38. def test_paraformer(self):
  39. inference_pipeline = pipeline(
  40. task=Tasks.auto_speech_recognition,
  41. model='damo/speech_data2vec_pretrain-paraformer-zh-cn-aishell2-16k')
  42. rec_result = inference_pipeline(
  43. audio_in='https://modelscope.oss-cn-beijing.aliyuncs.com/test/audios/asr_example.wav')
  44. logger.info("asr inference result: {0}".format(rec_result))
  45. assert rec_result["text"] == "每一天都要快乐喔"
  46. class TestMfccaInferencePipelines(unittest.TestCase):
  47. def test_funasr_path(self):
  48. import funasr
  49. import os
  50. logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
  51. def test_alimeeting(self):
  52. inference_pipeline = pipeline(
  53. task=Tasks.auto_speech_recognition,
  54. model='NPU-ASLP/speech_mfcca_asr-zh-cn-16k-alimeeting-vocab4950',
  55. model_revision='v3.0.0')
  56. rec_result = inference_pipeline(
  57. audio_in='https://pre.modelscope.cn/api/v1/models/NPU-ASLP/speech_mfcca_asr-zh-cn-16k-alimeeting-vocab4950/repo?Revision=master&FilePath=example/asr_example_mc.wav')
  58. logger.info("asr inference result: {0}".format(rec_result))
  59. class TestParaformerInferencePipelines(unittest.TestCase):
  60. def test_funasr_path(self):
  61. import funasr
  62. import os
  63. logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
  64. def test_paraformer_large_contextual_common(self):
  65. param_dict = dict()
  66. param_dict['hotword'] = "https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/hotword.txt"
  67. inference_pipeline = pipeline(
  68. task=Tasks.auto_speech_recognition,
  69. model='damo/speech_paraformer-large-contextual_asr_nat-zh-cn-16k-common-vocab8404',
  70. param_dict=param_dict)
  71. rec_result = inference_pipeline(
  72. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_hotword.wav')
  73. logger.info("asr inference result: {0}".format(rec_result))
  74. def test_paraformer_large_aishell1(self):
  75. inference_pipeline = pipeline(
  76. task=Tasks.auto_speech_recognition,
  77. model='damo/speech_paraformer-large_asr_nat-zh-cn-16k-aishell1-vocab8404-pytorch')
  78. rec_result = inference_pipeline(
  79. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  80. logger.info("asr inference result: {0}".format(rec_result))
  81. def test_paraformer_large_aishell2(self):
  82. inference_pipeline = pipeline(
  83. task=Tasks.auto_speech_recognition,
  84. model='damo/speech_paraformer-large_asr_nat-zh-cn-16k-aishell2-vocab8404-pytorch')
  85. rec_result = inference_pipeline(
  86. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  87. logger.info("asr inference result: {0}".format(rec_result))
  88. def test_paraformer_large_common(self):
  89. inference_pipeline = pipeline(
  90. task=Tasks.auto_speech_recognition,
  91. model='damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch')
  92. rec_result = inference_pipeline(
  93. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  94. logger.info("asr inference result: {0}".format(rec_result))
  95. def test_paraformer_large_online_common(self):
  96. inference_pipeline = pipeline(
  97. task=Tasks.auto_speech_recognition,
  98. model='damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-online')
  99. rec_result = inference_pipeline(
  100. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  101. logger.info("asr inference result: {0}".format(rec_result))
  102. def test_paraformer_online_common(self):
  103. inference_pipeline = pipeline(
  104. task=Tasks.auto_speech_recognition,
  105. model='damo/speech_paraformer_asr_nat-zh-cn-16k-common-vocab8404-online')
  106. rec_result = inference_pipeline(
  107. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  108. logger.info("asr inference result: {0}".format(rec_result))
  109. def test_paraformer_tiny_commandword(self):
  110. inference_pipeline = pipeline(
  111. task=Tasks.auto_speech_recognition,
  112. model='damo/speech_paraformer-tiny-commandword_asr_nat-zh-cn-16k-vocab544-pytorch')
  113. rec_result = inference_pipeline(
  114. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh_command.wav')
  115. logger.info("asr inference result: {0}".format(rec_result))
  116. def test_paraformer_8k(self):
  117. inference_pipeline = pipeline(
  118. task=Tasks.auto_speech_recognition,
  119. model='damo/speech_paraformer_asr_nat-zh-cn-8k-common-vocab8358-tensorflow1')
  120. rec_result = inference_pipeline(
  121. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_8K.wav')
  122. logger.info("asr inference result: {0}".format(rec_result))
  123. def test_paraformer_aishell1(self):
  124. inference_pipeline = pipeline(
  125. task=Tasks.auto_speech_recognition,
  126. model='damo/speech_paraformer_asr_nat-zh-cn-16k-aishell1-vocab4234-pytorch')
  127. rec_result = inference_pipeline(
  128. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  129. logger.info("asr inference result: {0}".format(rec_result))
  130. def test_paraformer_aishell2(self):
  131. inference_pipeline = pipeline(
  132. task=Tasks.auto_speech_recognition,
  133. model='damo/speech_paraformer_asr_nat-zh-cn-16k-aishell2-vocab5212-pytorch')
  134. rec_result = inference_pipeline(
  135. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  136. logger.info("asr inference result: {0}".format(rec_result))
  137. class TestParaformerBertInferencePipelines(unittest.TestCase):
  138. def test_funasr_path(self):
  139. import funasr
  140. import os
  141. logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
  142. def test_aishell1(self):
  143. inference_pipeline = pipeline(
  144. task=Tasks.auto_speech_recognition,
  145. model='damo/speech_paraformerbert_asr_nat-zh-cn-16k-aishell1-vocab4234-pytorch')
  146. rec_result = inference_pipeline(
  147. audio_in='https://modelscope.oss-cn-beijing.aliyuncs.com/test/audios/asr_example.wav')
  148. logger.info("asr inference result: {0}".format(rec_result))
  149. def test_aishell2(self):
  150. inference_pipeline = pipeline(
  151. task=Tasks.auto_speech_recognition,
  152. model='damo/speech_paraformerbert_asr_nat-zh-cn-16k-aishell2-vocab5212-pytorch')
  153. rec_result = inference_pipeline(
  154. audio_in='https://modelscope.oss-cn-beijing.aliyuncs.com/test/audios/asr_example.wav')
  155. logger.info("asr inference result: {0}".format(rec_result))
  156. class TestUniasrInferencePipelines(unittest.TestCase):
  157. def test_funasr_path(self):
  158. import funasr
  159. import os
  160. logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
  161. def test_uniasr_2pass_cantonese_chs_16k_common_offline(self):
  162. inference_pipeline = pipeline(
  163. task=Tasks.auto_speech_recognition,
  164. model='damo/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-online')
  165. rec_result = inference_pipeline(
  166. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_cantonese-CHS.wav',
  167. param_dict={"decoding_model": "offline"})
  168. logger.info("asr inference result: {0}".format(rec_result))
  169. def test_uniasr_2pass_cantonese_chs_16k_common_online(self):
  170. inference_pipeline = pipeline(
  171. task=Tasks.auto_speech_recognition,
  172. model='damo/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-online')
  173. rec_result = inference_pipeline(
  174. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_cantonese-CHS.wav',
  175. param_dict={"decoding_model": "normal"})
  176. logger.info("asr inference result: {0}".format(rec_result))
  177. def test_uniasr_2pass_cn_dialect_offline(self):
  178. inference_pipeline = pipeline(
  179. task=Tasks.auto_speech_recognition,
  180. model='damo/speech_UniASR_asr_2pass-cn-dialect-16k-vocab8358-tensorflow1-offline')
  181. rec_result = inference_pipeline(
  182. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  183. logger.info("asr inference result: {0}".format(rec_result))
  184. def test_uniasr_2pass_cn_dialect_online(self):
  185. inference_pipeline = pipeline(
  186. task=Tasks.auto_speech_recognition,
  187. model='damo/speech_UniASR_asr_2pass-cn-dialect-16k-vocab8358-tensorflow1-online')
  188. rec_result = inference_pipeline(
  189. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  190. logger.info("asr inference result: {0}".format(rec_result))
  191. def test_uniasr_2pass_de_common_offline(self):
  192. inference_pipeline = pipeline(
  193. task=Tasks.auto_speech_recognition,
  194. model='damo/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-offline')
  195. rec_result = inference_pipeline(
  196. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_de.wav',
  197. param_dict={"decoding_model": "offline"})
  198. logger.info("asr inference result: {0}".format(rec_result))
  199. def test_uniasr_2pass_de_common_online(self):
  200. inference_pipeline = pipeline(
  201. task=Tasks.auto_speech_recognition,
  202. model='damo/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-online')
  203. rec_result = inference_pipeline(
  204. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_de.wav',
  205. param_dict={"decoding_model": "normal"})
  206. logger.info("asr inference result: {0}".format(rec_result))
  207. def test_uniasr_2pass_en_common_offline(self):
  208. inference_pipeline = pipeline(
  209. task=Tasks.auto_speech_recognition,
  210. model='damo/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-offline')
  211. rec_result = inference_pipeline(
  212. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_en.wav',
  213. param_dict={"decoding_model": "offline"})
  214. logger.info("asr inference result: {0}".format(rec_result))
  215. def test_uniasr_2pass_en_common_online(self):
  216. inference_pipeline = pipeline(
  217. task=Tasks.auto_speech_recognition,
  218. model='damo/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-online')
  219. rec_result = inference_pipeline(
  220. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_en.wav',
  221. param_dict={"decoding_model": "normal"})
  222. logger.info("asr inference result: {0}".format(rec_result))
  223. def test_uniasr_2pass_es_common_offline(self):
  224. inference_pipeline = pipeline(
  225. task=Tasks.auto_speech_recognition,
  226. model='damo/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-offline')
  227. rec_result = inference_pipeline(
  228. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_es.wav',
  229. param_dict={"decoding_model": "offline"})
  230. logger.info("asr inference result: {0}".format(rec_result))
  231. def test_uniasr_2pass_es_common_online(self):
  232. inference_pipeline = pipeline(
  233. task=Tasks.auto_speech_recognition,
  234. model='damo/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-online')
  235. rec_result = inference_pipeline(
  236. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_es.wav',
  237. param_dict={"decoding_model": "normal"})
  238. logger.info("asr inference result: {0}".format(rec_result))
  239. def test_uniasr_2pass_fa_common_offline(self):
  240. inference_pipeline = pipeline(
  241. task=Tasks.auto_speech_recognition,
  242. model='damo/speech_UniASR_asr_2pass-fa-16k-common-vocab1257-pytorch-offline')
  243. rec_result = inference_pipeline(
  244. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_fa.wav',
  245. param_dict={"decoding_model": "offline"})
  246. logger.info("asr inference result: {0}".format(rec_result))
  247. def test_uniasr_2pass_fa_common_online(self):
  248. inference_pipeline = pipeline(
  249. task=Tasks.auto_speech_recognition,
  250. model='damo/speech_UniASR_asr_2pass-fa-16k-common-vocab1257-pytorch-online')
  251. rec_result = inference_pipeline(
  252. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_fa.wav',
  253. param_dict={"decoding_model": "normal"})
  254. logger.info("asr inference result: {0}".format(rec_result))
  255. def test_uniasr_2pass_fr_common_offline(self):
  256. inference_pipeline = pipeline(
  257. task=Tasks.auto_speech_recognition,
  258. model='damo/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-offline')
  259. rec_result = inference_pipeline(
  260. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_fr.wav',
  261. param_dict={"decoding_model": "offline"})
  262. logger.info("asr inference result: {0}".format(rec_result))
  263. def test_uniasr_2pass_fr_common_online(self):
  264. inference_pipeline = pipeline(
  265. task=Tasks.auto_speech_recognition,
  266. model='damo/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-online')
  267. rec_result = inference_pipeline(
  268. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_fr.wav',
  269. param_dict={"decoding_model": "normal"})
  270. logger.info("asr inference result: {0}".format(rec_result))
  271. def test_uniasr_2pass_id_common_offline(self):
  272. inference_pipeline = pipeline(
  273. task=Tasks.auto_speech_recognition,
  274. model='damo/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-online')
  275. rec_result = inference_pipeline(
  276. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_id.wav',
  277. param_dict={"decoding_model": "offline"})
  278. logger.info("asr inference result: {0}".format(rec_result))
  279. def test_uniasr_2pass_id_common_online(self):
  280. inference_pipeline = pipeline(
  281. task=Tasks.auto_speech_recognition,
  282. model='damo/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-online')
  283. rec_result = inference_pipeline(
  284. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_id.wav',
  285. param_dict={"decoding_model": "normal"})
  286. logger.info("asr inference result: {0}".format(rec_result))
  287. def test_uniasr_2pass_ja_common_offline(self):
  288. inference_pipeline = pipeline(
  289. task=Tasks.auto_speech_recognition,
  290. model='damo/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-offline')
  291. rec_result = inference_pipeline(
  292. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_ja.wav',
  293. param_dict={"decoding_model": "offline"})
  294. logger.info("asr inference result: {0}".format(rec_result))
  295. def test_uniasr_2pass_ja_common_online(self):
  296. inference_pipeline = pipeline(
  297. task=Tasks.auto_speech_recognition,
  298. model='damo/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-online')
  299. rec_result = inference_pipeline(
  300. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_ja.wav',
  301. param_dict={"decoding_model": "normal"})
  302. logger.info("asr inference result: {0}".format(rec_result))
  303. def test_uniasr_2pass_ko_common_offline(self):
  304. inference_pipeline = pipeline(
  305. task=Tasks.auto_speech_recognition,
  306. model='damo/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-offline')
  307. rec_result = inference_pipeline(
  308. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_ko.wav',
  309. param_dict={"decoding_model": "offline"})
  310. logger.info("asr inference result: {0}".format(rec_result))
  311. def test_uniasr_2pass_ko_common_online(self):
  312. inference_pipeline = pipeline(
  313. task=Tasks.auto_speech_recognition,
  314. model='damo/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-online')
  315. rec_result = inference_pipeline(
  316. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_ko.wav',
  317. param_dict={"decoding_model": "normal"})
  318. logger.info("asr inference result: {0}".format(rec_result))
  319. def test_uniasr_2pass_minnan_common_offline(self):
  320. inference_pipeline = pipeline(
  321. task=Tasks.auto_speech_recognition,
  322. model='damo/speech_UniASR_asr_2pass-minnan-16k-common-vocab3825')
  323. rec_result = inference_pipeline(
  324. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav',
  325. param_dict={"decoding_model": "offline"})
  326. logger.info("asr inference result: {0}".format(rec_result))
  327. def test_uniasr_2pass_pt_common_offline(self):
  328. inference_pipeline = pipeline(
  329. task=Tasks.auto_speech_recognition,
  330. model='damo/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-offline')
  331. rec_result = inference_pipeline(
  332. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_pt.wav',
  333. param_dict={"decoding_model": "offline"})
  334. logger.info("asr inference result: {0}".format(rec_result))
  335. def test_uniasr_2pass_pt_common_online(self):
  336. inference_pipeline = pipeline(
  337. task=Tasks.auto_speech_recognition,
  338. model='damo/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-online')
  339. rec_result = inference_pipeline(
  340. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_pt.wav',
  341. param_dict={"decoding_model": "normal"})
  342. logger.info("asr inference result: {0}".format(rec_result))
  343. def test_uniasr_2pass_ru_common_offline(self):
  344. inference_pipeline = pipeline(
  345. task=Tasks.auto_speech_recognition,
  346. model='damo/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-offline')
  347. rec_result = inference_pipeline(
  348. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_ru.wav',
  349. param_dict={"decoding_model": "offline"})
  350. logger.info("asr inference result: {0}".format(rec_result))
  351. def test_uniasr_2pass_ru_common_online(self):
  352. inference_pipeline = pipeline(
  353. task=Tasks.auto_speech_recognition,
  354. model='damo/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-online')
  355. rec_result = inference_pipeline(
  356. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_ru.wav',
  357. param_dict={"decoding_model": "normal"})
  358. logger.info("asr inference result: {0}".format(rec_result))
  359. def test_uniasr_2pass_vi_common_offline(self):
  360. inference_pipeline = pipeline(
  361. task=Tasks.auto_speech_recognition,
  362. model='damo/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-offline')
  363. rec_result = inference_pipeline(
  364. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_vi.wav',
  365. param_dict={"decoding_model": "offline"})
  366. logger.info("asr inference result: {0}".format(rec_result))
  367. def test_uniasr_2pass_vi_common_online(self):
  368. inference_pipeline = pipeline(
  369. task=Tasks.auto_speech_recognition,
  370. model='damo/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-online')
  371. rec_result = inference_pipeline(
  372. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_vi.wav',
  373. param_dict={"decoding_model": "normal"})
  374. logger.info("asr inference result: {0}".format(rec_result))
  375. def test_uniasr_2pass_zhcn_8k_common_vocab3445_offline(self):
  376. inference_pipeline = pipeline(
  377. task=Tasks.auto_speech_recognition,
  378. model='damo/speech_UniASR_asr_2pass-zh-cn-8k-common-vocab3445-pytorch-offline')
  379. rec_result = inference_pipeline(
  380. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav',
  381. param_dict={"decoding_model": "offline"})
  382. logger.info("asr inference result: {0}".format(rec_result))
  383. def test_uniasr_2pass_zhcn_8k_common_vocab3445_online(self):
  384. inference_pipeline = pipeline(
  385. task=Tasks.auto_speech_recognition,
  386. model='damo/speech_UniASR_asr_2pass-zh-cn-8k-common-vocab3445-pytorch-online')
  387. rec_result = inference_pipeline(
  388. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav',
  389. param_dict={"decoding_model": "normal"})
  390. logger.info("asr inference result: {0}".format(rec_result))
  391. def test_uniasr_2pass_zhcn_8k_common_vocab8358_offline(self):
  392. inference_pipeline = pipeline(
  393. task=Tasks.auto_speech_recognition,
  394. model='damo/speech_UniASR_asr_2pass-zh-cn-8k-common-vocab8358-tensorflow1-offline')
  395. rec_result = inference_pipeline(
  396. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav',
  397. param_dict={"decoding_model": "offline"})
  398. logger.info("asr inference result: {0}".format(rec_result))
  399. def test_uniasr_2pass_zhcn_8k_common_vocab8358_online(self):
  400. inference_pipeline = pipeline(
  401. task=Tasks.auto_speech_recognition,
  402. model='damo/speech_UniASR_asr_2pass-zh-cn-8k-common-vocab8358-tensorflow1-online')
  403. rec_result = inference_pipeline(
  404. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav',
  405. param_dict={"decoding_model": "normal"})
  406. logger.info("asr inference result: {0}".format(rec_result))
  407. def test_uniasr_2pass_zhcn_16k_common_vocab8358_offline(self):
  408. inference_pipeline = pipeline(
  409. task=Tasks.auto_speech_recognition,
  410. model='damo/speech_UniASR_asr_2pass-zh-cn-16k-common-vocab8358-tensorflow1-offline')
  411. rec_result = inference_pipeline(
  412. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav',
  413. param_dict={"decoding_model": "offline"})
  414. logger.info("asr inference result: {0}".format(rec_result))
  415. def test_uniasr_2pass_zhcn_16k_common_vocab8358_online(self):
  416. inference_pipeline = pipeline(
  417. task=Tasks.auto_speech_recognition,
  418. model='damo/speech_UniASR_asr_2pass-zh-cn-16k-common-vocab8358-tensorflow1-online')
  419. rec_result = inference_pipeline(
  420. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav',
  421. param_dict={"decoding_model": "normal"})
  422. logger.info("asr inference result: {0}".format(rec_result))
  423. if __name__ == '__main__':
  424. unittest.main()