demo_funasr.py 670 B

12345678910111213141516171819
  1. #!/usr/bin/env python3
  2. # -*- encoding: utf-8 -*-
  3. # Copyright FunASR (https://github.com/alibaba-damo-academy/FunASR). All Rights Reserved.
  4. # MIT License (https://opensource.org/licenses/MIT)
  5. from funasr import AutoModel
  6. multilingual_wavs = [
  7. "example_zh-CN.mp3",
  8. "example_en.mp3",
  9. "example_ja.mp3",
  10. "example_ko.mp3",
  11. ]
  12. model = AutoModel(model="iic/speech_whisper-large_lid_multilingual_pytorch", model_revision="v2.0.4")
  13. for wav_id in multilingual_wavs:
  14. wav_file = f"{model.model_path}/examples/{wav_id}"
  15. res = model.generate(input=wav_file, data_type="sound", inference_clip_length=250)
  16. print("detect sample {}: {}".format(wav_id, res))