test_asr_inference_pipeline.py 26 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505
  1. import unittest
  2. from modelscope.pipelines import pipeline
  3. from modelscope.utils.constant import Tasks
  4. from modelscope.utils.logger import get_logger
  5. logger = get_logger()
  6. class TestConformerInferencePipelines(unittest.TestCase):
  7. def test_funasr_path(self):
  8. import funasr
  9. import os
  10. logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
  11. def test_aishell1(self):
  12. inference_pipeline = pipeline(
  13. task=Tasks.auto_speech_recognition,
  14. model='damo/speech_conformer_asr_nat-zh-cn-16k-aishell1-vocab4234-pytorch')
  15. rec_result = inference_pipeline(
  16. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  17. logger.info("asr inference result: {0}".format(rec_result))
  18. def test_aishell2(self):
  19. inference_pipeline = pipeline(
  20. task=Tasks.auto_speech_recognition,
  21. model='damo/speech_conformer_asr_nat-zh-cn-16k-aishell2-vocab5212-pytorch')
  22. rec_result = inference_pipeline(
  23. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  24. logger.info("asr inference result: {0}".format(rec_result))
  25. class TestData2vecInferencePipelines(unittest.TestCase):
  26. def test_funasr_path(self):
  27. import funasr
  28. import os
  29. logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
  30. def test_transformer(self):
  31. inference_pipeline = pipeline(
  32. task=Tasks.auto_speech_recognition,
  33. model='damo/speech_data2vec_pretrain-zh-cn-aishell2-16k-pytorch')
  34. rec_result = inference_pipeline(
  35. audio_in='https://modelscope.oss-cn-beijing.aliyuncs.com/test/audios/asr_example.wav')
  36. logger.info("asr inference result: {0}".format(rec_result))
  37. assert rec_result["text"] == "每一天都要快乐喔"
  38. def test_paraformer(self):
  39. inference_pipeline = pipeline(
  40. task=Tasks.auto_speech_recognition,
  41. model='damo/speech_data2vec_pretrain-paraformer-zh-cn-aishell2-16k')
  42. rec_result = inference_pipeline(
  43. audio_in='https://modelscope.oss-cn-beijing.aliyuncs.com/test/audios/asr_example.wav')
  44. logger.info("asr inference result: {0}".format(rec_result))
  45. assert rec_result["text"] == "每一天都要快乐喔"
  46. class TestMfccaInferencePipelines(unittest.TestCase):
  47. def test_funasr_path(self):
  48. import funasr
  49. import os
  50. logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
  51. def test_alimeeting(self):
  52. inference_pipeline = pipeline(
  53. task=Tasks.auto_speech_recognition,
  54. model='NPU-ASLP/speech_mfcca_asr-zh-cn-16k-alimeeting-vocab4950',
  55. model_revision='v3.0.0')
  56. rec_result = inference_pipeline(
  57. audio_in='https://pre.modelscope.cn/api/v1/models/NPU-ASLP/speech_mfcca_asr-zh-cn-16k-alimeeting-vocab4950/repo?Revision=master&FilePath=example/asr_example_mc.wav')
  58. logger.info("asr inference result: {0}".format(rec_result))
  59. class TestParaformerInferencePipelines(unittest.TestCase):
  60. def test_funasr_path(self):
  61. import funasr
  62. import os
  63. logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
  64. def test_paraformer_large_contextual_common(self):
  65. param_dict = dict()
  66. param_dict['hotword'] = "https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/hotword.txt"
  67. inference_pipeline = pipeline(
  68. task=Tasks.auto_speech_recognition,
  69. model='damo/speech_paraformer-large-contextual_asr_nat-zh-cn-16k-common-vocab8404',
  70. param_dict=param_dict)
  71. rec_result = inference_pipeline(
  72. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_hotword.wav')
  73. logger.info("asr inference result: {0}".format(rec_result))
  74. assert rec_result["text"] == "国务院发展研究中心市场经济研究所副所长邓郁松认为"
  75. def test_paraformer_large_aishell1(self):
  76. inference_pipeline = pipeline(
  77. task=Tasks.auto_speech_recognition,
  78. model='damo/speech_paraformer-large_asr_nat-zh-cn-16k-aishell1-vocab8404-pytorch')
  79. rec_result = inference_pipeline(
  80. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  81. logger.info("asr inference result: {0}".format(rec_result))
  82. assert rec_result["text"] == "欢迎大家来体验达摩院推出的语音识别模型"
  83. def test_paraformer_large_aishell2(self):
  84. inference_pipeline = pipeline(
  85. task=Tasks.auto_speech_recognition,
  86. model='damo/speech_paraformer-large_asr_nat-zh-cn-16k-aishell2-vocab8404-pytorch')
  87. rec_result = inference_pipeline(
  88. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  89. logger.info("asr inference result: {0}".format(rec_result))
  90. assert rec_result["text"] == "欢迎大家来体验达摩院推出的语音识别模型"
  91. def test_paraformer_large_common(self):
  92. inference_pipeline = pipeline(
  93. task=Tasks.auto_speech_recognition,
  94. model='damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch')
  95. rec_result = inference_pipeline(
  96. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  97. logger.info("asr inference result: {0}".format(rec_result))
  98. assert rec_result["text"] == "欢迎大家来体验达摩院推出的语音识别模型"
  99. def test_paraformer_large_online_common(self):
  100. inference_pipeline = pipeline(
  101. task=Tasks.auto_speech_recognition,
  102. model='damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-online',
  103. model_revision='v1.0.6',
  104. update_model=False,
  105. mode="paraformer_fake_streaming"
  106. )
  107. rec_result = inference_pipeline(
  108. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  109. logger.info("asr inference result: {0}".format(rec_result))
  110. assert rec_result["text"] == "欢迎大家来体验达摩院推出的语音识别模型"
  111. def test_paraformer_online_common(self):
  112. inference_pipeline = pipeline(
  113. task=Tasks.auto_speech_recognition,
  114. model='damo/speech_paraformer_asr_nat-zh-cn-16k-common-vocab8404-online',
  115. model_revision='v1.0.6',
  116. update_model=False,
  117. mode="paraformer_fake_streaming"
  118. )
  119. rec_result = inference_pipeline(
  120. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  121. logger.info("asr inference result: {0}".format(rec_result))
  122. assert rec_result["text"] == "欢迎大家来体验达摩院推出的语音识别模型"
  123. def test_paraformer_tiny_commandword(self):
  124. inference_pipeline = pipeline(
  125. task=Tasks.auto_speech_recognition,
  126. model='damo/speech_paraformer-tiny-commandword_asr_nat-zh-cn-16k-vocab544-pytorch')
  127. rec_result = inference_pipeline(
  128. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh_command.wav')
  129. logger.info("asr inference result: {0}".format(rec_result))
  130. def test_paraformer_8k(self):
  131. inference_pipeline = pipeline(
  132. task=Tasks.auto_speech_recognition,
  133. model='damo/speech_paraformer_asr_nat-zh-cn-8k-common-vocab8358-tensorflow1')
  134. rec_result = inference_pipeline(
  135. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_8K.wav')
  136. logger.info("asr inference result: {0}".format(rec_result))
  137. def test_paraformer_aishell1(self):
  138. inference_pipeline = pipeline(
  139. task=Tasks.auto_speech_recognition,
  140. model='damo/speech_paraformer_asr_nat-zh-cn-16k-aishell1-vocab4234-pytorch')
  141. rec_result = inference_pipeline(
  142. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  143. logger.info("asr inference result: {0}".format(rec_result))
  144. def test_paraformer_aishell2(self):
  145. inference_pipeline = pipeline(
  146. task=Tasks.auto_speech_recognition,
  147. model='damo/speech_paraformer_asr_nat-zh-cn-16k-aishell2-vocab5212-pytorch')
  148. rec_result = inference_pipeline(
  149. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  150. logger.info("asr inference result: {0}".format(rec_result))
  151. class TestParaformerBertInferencePipelines(unittest.TestCase):
  152. def test_funasr_path(self):
  153. import funasr
  154. import os
  155. logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
  156. def test_aishell1(self):
  157. inference_pipeline = pipeline(
  158. task=Tasks.auto_speech_recognition,
  159. model='damo/speech_paraformerbert_asr_nat-zh-cn-16k-aishell1-vocab4234-pytorch')
  160. rec_result = inference_pipeline(
  161. audio_in='https://modelscope.oss-cn-beijing.aliyuncs.com/test/audios/asr_example.wav')
  162. logger.info("asr inference result: {0}".format(rec_result))
  163. def test_aishell2(self):
  164. inference_pipeline = pipeline(
  165. task=Tasks.auto_speech_recognition,
  166. model='damo/speech_paraformerbert_asr_nat-zh-cn-16k-aishell2-vocab5212-pytorch')
  167. rec_result = inference_pipeline(
  168. audio_in='https://modelscope.oss-cn-beijing.aliyuncs.com/test/audios/asr_example.wav')
  169. logger.info("asr inference result: {0}".format(rec_result))
  170. class TestUniasrInferencePipelines(unittest.TestCase):
  171. def test_funasr_path(self):
  172. import funasr
  173. import os
  174. logger.info("run_dir:{0} ; funasr_path: {1}".format(os.getcwd(), funasr.__file__))
  175. def test_uniasr_2pass_cantonese_chs_16k_common_offline(self):
  176. inference_pipeline = pipeline(
  177. task=Tasks.auto_speech_recognition,
  178. model='damo/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-online')
  179. rec_result = inference_pipeline(
  180. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_cantonese-CHS.wav',
  181. param_dict={"decoding_model": "offline"})
  182. logger.info("asr inference result: {0}".format(rec_result))
  183. def test_uniasr_2pass_cantonese_chs_16k_common_online(self):
  184. inference_pipeline = pipeline(
  185. task=Tasks.auto_speech_recognition,
  186. model='damo/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-online')
  187. rec_result = inference_pipeline(
  188. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_cantonese-CHS.wav',
  189. param_dict={"decoding_model": "normal"})
  190. logger.info("asr inference result: {0}".format(rec_result))
  191. def test_uniasr_2pass_cn_dialect_offline(self):
  192. inference_pipeline = pipeline(
  193. task=Tasks.auto_speech_recognition,
  194. model='damo/speech_UniASR_asr_2pass-cn-dialect-16k-vocab8358-tensorflow1-offline')
  195. rec_result = inference_pipeline(
  196. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  197. logger.info("asr inference result: {0}".format(rec_result))
  198. def test_uniasr_2pass_cn_dialect_online(self):
  199. inference_pipeline = pipeline(
  200. task=Tasks.auto_speech_recognition,
  201. model='damo/speech_UniASR_asr_2pass-cn-dialect-16k-vocab8358-tensorflow1-online')
  202. rec_result = inference_pipeline(
  203. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav')
  204. logger.info("asr inference result: {0}".format(rec_result))
  205. def test_uniasr_2pass_de_common_offline(self):
  206. inference_pipeline = pipeline(
  207. task=Tasks.auto_speech_recognition,
  208. model='damo/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-offline')
  209. rec_result = inference_pipeline(
  210. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_de.wav',
  211. param_dict={"decoding_model": "offline"})
  212. logger.info("asr inference result: {0}".format(rec_result))
  213. def test_uniasr_2pass_de_common_online(self):
  214. inference_pipeline = pipeline(
  215. task=Tasks.auto_speech_recognition,
  216. model='damo/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-online')
  217. rec_result = inference_pipeline(
  218. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_de.wav',
  219. param_dict={"decoding_model": "normal"})
  220. logger.info("asr inference result: {0}".format(rec_result))
  221. def test_uniasr_2pass_en_common_offline(self):
  222. inference_pipeline = pipeline(
  223. task=Tasks.auto_speech_recognition,
  224. model='damo/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-offline')
  225. rec_result = inference_pipeline(
  226. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_en.wav',
  227. param_dict={"decoding_model": "offline"})
  228. logger.info("asr inference result: {0}".format(rec_result))
  229. def test_uniasr_2pass_en_common_online(self):
  230. inference_pipeline = pipeline(
  231. task=Tasks.auto_speech_recognition,
  232. model='damo/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-online')
  233. rec_result = inference_pipeline(
  234. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_en.wav',
  235. param_dict={"decoding_model": "normal"})
  236. logger.info("asr inference result: {0}".format(rec_result))
  237. def test_uniasr_2pass_es_common_offline(self):
  238. inference_pipeline = pipeline(
  239. task=Tasks.auto_speech_recognition,
  240. model='damo/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-offline')
  241. rec_result = inference_pipeline(
  242. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_es.wav',
  243. param_dict={"decoding_model": "offline"})
  244. logger.info("asr inference result: {0}".format(rec_result))
  245. def test_uniasr_2pass_es_common_online(self):
  246. inference_pipeline = pipeline(
  247. task=Tasks.auto_speech_recognition,
  248. model='damo/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-online')
  249. rec_result = inference_pipeline(
  250. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_es.wav',
  251. param_dict={"decoding_model": "normal"})
  252. logger.info("asr inference result: {0}".format(rec_result))
  253. def test_uniasr_2pass_fa_common_offline(self):
  254. inference_pipeline = pipeline(
  255. task=Tasks.auto_speech_recognition,
  256. model='damo/speech_UniASR_asr_2pass-fa-16k-common-vocab1257-pytorch-offline')
  257. rec_result = inference_pipeline(
  258. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_fa.wav',
  259. param_dict={"decoding_model": "offline"})
  260. logger.info("asr inference result: {0}".format(rec_result))
  261. def test_uniasr_2pass_fa_common_online(self):
  262. inference_pipeline = pipeline(
  263. task=Tasks.auto_speech_recognition,
  264. model='damo/speech_UniASR_asr_2pass-fa-16k-common-vocab1257-pytorch-online')
  265. rec_result = inference_pipeline(
  266. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_fa.wav',
  267. param_dict={"decoding_model": "normal"})
  268. logger.info("asr inference result: {0}".format(rec_result))
  269. def test_uniasr_2pass_fr_common_offline(self):
  270. inference_pipeline = pipeline(
  271. task=Tasks.auto_speech_recognition,
  272. model='damo/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-offline')
  273. rec_result = inference_pipeline(
  274. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_fr.wav',
  275. param_dict={"decoding_model": "offline"})
  276. logger.info("asr inference result: {0}".format(rec_result))
  277. def test_uniasr_2pass_fr_common_online(self):
  278. inference_pipeline = pipeline(
  279. task=Tasks.auto_speech_recognition,
  280. model='damo/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-online')
  281. rec_result = inference_pipeline(
  282. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_fr.wav',
  283. param_dict={"decoding_model": "normal"})
  284. logger.info("asr inference result: {0}".format(rec_result))
  285. def test_uniasr_2pass_id_common_offline(self):
  286. inference_pipeline = pipeline(
  287. task=Tasks.auto_speech_recognition,
  288. model='damo/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-online')
  289. rec_result = inference_pipeline(
  290. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_id.wav',
  291. param_dict={"decoding_model": "offline"})
  292. logger.info("asr inference result: {0}".format(rec_result))
  293. def test_uniasr_2pass_id_common_online(self):
  294. inference_pipeline = pipeline(
  295. task=Tasks.auto_speech_recognition,
  296. model='damo/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-online')
  297. rec_result = inference_pipeline(
  298. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_id.wav',
  299. param_dict={"decoding_model": "normal"})
  300. logger.info("asr inference result: {0}".format(rec_result))
  301. def test_uniasr_2pass_ja_common_offline(self):
  302. inference_pipeline = pipeline(
  303. task=Tasks.auto_speech_recognition,
  304. model='damo/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-offline')
  305. rec_result = inference_pipeline(
  306. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_ja.wav',
  307. param_dict={"decoding_model": "offline"})
  308. logger.info("asr inference result: {0}".format(rec_result))
  309. def test_uniasr_2pass_ja_common_online(self):
  310. inference_pipeline = pipeline(
  311. task=Tasks.auto_speech_recognition,
  312. model='damo/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-online')
  313. rec_result = inference_pipeline(
  314. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_ja.wav',
  315. param_dict={"decoding_model": "normal"})
  316. logger.info("asr inference result: {0}".format(rec_result))
  317. def test_uniasr_2pass_ko_common_offline(self):
  318. inference_pipeline = pipeline(
  319. task=Tasks.auto_speech_recognition,
  320. model='damo/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-offline')
  321. rec_result = inference_pipeline(
  322. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_ko.wav',
  323. param_dict={"decoding_model": "offline"})
  324. logger.info("asr inference result: {0}".format(rec_result))
  325. def test_uniasr_2pass_ko_common_online(self):
  326. inference_pipeline = pipeline(
  327. task=Tasks.auto_speech_recognition,
  328. model='damo/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-online')
  329. rec_result = inference_pipeline(
  330. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_ko.wav',
  331. param_dict={"decoding_model": "normal"})
  332. logger.info("asr inference result: {0}".format(rec_result))
  333. def test_uniasr_2pass_minnan_common_offline(self):
  334. inference_pipeline = pipeline(
  335. task=Tasks.auto_speech_recognition,
  336. model='damo/speech_UniASR_asr_2pass-minnan-16k-common-vocab3825')
  337. rec_result = inference_pipeline(
  338. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav',
  339. param_dict={"decoding_model": "offline"})
  340. logger.info("asr inference result: {0}".format(rec_result))
  341. def test_uniasr_2pass_pt_common_offline(self):
  342. inference_pipeline = pipeline(
  343. task=Tasks.auto_speech_recognition,
  344. model='damo/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-offline')
  345. rec_result = inference_pipeline(
  346. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_pt.wav',
  347. param_dict={"decoding_model": "offline"})
  348. logger.info("asr inference result: {0}".format(rec_result))
  349. def test_uniasr_2pass_pt_common_online(self):
  350. inference_pipeline = pipeline(
  351. task=Tasks.auto_speech_recognition,
  352. model='damo/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-online')
  353. rec_result = inference_pipeline(
  354. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_pt.wav',
  355. param_dict={"decoding_model": "normal"})
  356. logger.info("asr inference result: {0}".format(rec_result))
  357. def test_uniasr_2pass_ru_common_offline(self):
  358. inference_pipeline = pipeline(
  359. task=Tasks.auto_speech_recognition,
  360. model='damo/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-offline')
  361. rec_result = inference_pipeline(
  362. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_ru.wav',
  363. param_dict={"decoding_model": "offline"})
  364. logger.info("asr inference result: {0}".format(rec_result))
  365. def test_uniasr_2pass_ru_common_online(self):
  366. inference_pipeline = pipeline(
  367. task=Tasks.auto_speech_recognition,
  368. model='damo/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-online')
  369. rec_result = inference_pipeline(
  370. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_ru.wav',
  371. param_dict={"decoding_model": "normal"})
  372. logger.info("asr inference result: {0}".format(rec_result))
  373. def test_uniasr_2pass_vi_common_offline(self):
  374. inference_pipeline = pipeline(
  375. task=Tasks.auto_speech_recognition,
  376. model='damo/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-offline')
  377. rec_result = inference_pipeline(
  378. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_vi.wav',
  379. param_dict={"decoding_model": "offline"})
  380. logger.info("asr inference result: {0}".format(rec_result))
  381. def test_uniasr_2pass_vi_common_online(self):
  382. inference_pipeline = pipeline(
  383. task=Tasks.auto_speech_recognition,
  384. model='damo/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-online')
  385. rec_result = inference_pipeline(
  386. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_vi.wav',
  387. param_dict={"decoding_model": "normal"})
  388. logger.info("asr inference result: {0}".format(rec_result))
  389. def test_uniasr_2pass_zhcn_8k_common_vocab3445_offline(self):
  390. inference_pipeline = pipeline(
  391. task=Tasks.auto_speech_recognition,
  392. model='damo/speech_UniASR_asr_2pass-zh-cn-8k-common-vocab3445-pytorch-offline')
  393. rec_result = inference_pipeline(
  394. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav',
  395. param_dict={"decoding_model": "offline"})
  396. logger.info("asr inference result: {0}".format(rec_result))
  397. def test_uniasr_2pass_zhcn_8k_common_vocab3445_online(self):
  398. inference_pipeline = pipeline(
  399. task=Tasks.auto_speech_recognition,
  400. model='damo/speech_UniASR_asr_2pass-zh-cn-8k-common-vocab3445-pytorch-online')
  401. rec_result = inference_pipeline(
  402. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav',
  403. param_dict={"decoding_model": "normal"})
  404. logger.info("asr inference result: {0}".format(rec_result))
  405. def test_uniasr_2pass_zhcn_8k_common_vocab8358_offline(self):
  406. inference_pipeline = pipeline(
  407. task=Tasks.auto_speech_recognition,
  408. model='damo/speech_UniASR_asr_2pass-zh-cn-8k-common-vocab8358-tensorflow1-offline')
  409. rec_result = inference_pipeline(
  410. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav',
  411. param_dict={"decoding_model": "offline"})
  412. logger.info("asr inference result: {0}".format(rec_result))
  413. def test_uniasr_2pass_zhcn_8k_common_vocab8358_online(self):
  414. inference_pipeline = pipeline(
  415. task=Tasks.auto_speech_recognition,
  416. model='damo/speech_UniASR_asr_2pass-zh-cn-8k-common-vocab8358-tensorflow1-online')
  417. rec_result = inference_pipeline(
  418. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav',
  419. param_dict={"decoding_model": "normal"})
  420. logger.info("asr inference result: {0}".format(rec_result))
  421. def test_uniasr_2pass_zhcn_16k_common_vocab8358_offline(self):
  422. inference_pipeline = pipeline(
  423. task=Tasks.auto_speech_recognition,
  424. model='damo/speech_UniASR_asr_2pass-zh-cn-16k-common-vocab8358-tensorflow1-offline')
  425. rec_result = inference_pipeline(
  426. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav',
  427. param_dict={"decoding_model": "offline"})
  428. logger.info("asr inference result: {0}".format(rec_result))
  429. def test_uniasr_2pass_zhcn_16k_common_vocab8358_online(self):
  430. inference_pipeline = pipeline(
  431. task=Tasks.auto_speech_recognition,
  432. model='damo/speech_UniASR_asr_2pass-zh-cn-16k-common-vocab8358-tensorflow1-online')
  433. rec_result = inference_pipeline(
  434. audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav',
  435. param_dict={"decoding_model": "normal"})
  436. logger.info("asr inference result: {0}".format(rec_result))
  437. if __name__ == '__main__':
  438. unittest.main()