Sem descrição

hnluo 3aef6df36d Merge pull request #770 from alibaba-damo-academy/dev_lhn há 2 anos atrás
.github a43b3da4be update há 2 anos atrás
docs 2a31a37b7a readme docs há 2 anos atrás
egs 128caa57c2 update há 2 anos atrás
egs_modelscope 3aef6df36d Merge pull request #770 from alibaba-damo-academy/dev_lhn há 2 anos atrás
fun_text_processing 81a64b0e33 Update id_unit_test.tsv (#563) há 2 anos atrás
funasr 0063e4356a Merge pull request #768 from alibaba-damo-academy/dev_lhn há 2 anos atrás
tests edb7459e74 Update test_asr_inference_pipeline.py há 2 anos atrás
.gitignore c506a5f67e sdk utils há 2 anos atrás
Acknowledge 8545d6b203 docs zh há 2 anos atrás
MODEL_LICENSE 787b9d8e7e model license há 2 anos atrás
README.md 3709ad0f24 readme docs há 2 anos atrás
README_zh.md 93ad0b19c5 Update README_zh.md há 2 anos atrás
setup.py 4a2cb0c985 Update setup.py há 2 anos atrás

README.md

(简体中文|English)

FunASR: A Fundamental End-to-End Speech Recognition Toolkit

FunASR hopes to build a bridge between academic research and industrial applications on speech recognition. By supporting the training & finetuning of the industrial-grade speech recognition model released on ModelScope, researchers and developers can conduct research and production of speech recognition models more conveniently, and promote the development of speech recognition ecology. ASR for Fun!

News | Highlights | Installation | Quick Start | Runtime | Model Zoo | Contact

Highlights

  • FunASR is a fundamental speech recognition toolkit that offers a variety of features, including speech recognition (ASR), Voice Activity Detection (VAD), Punctuation Restoration, Language Models, Speaker Verification, Speaker Diarization and multi-talker ASR. FunASR provides convenient scripts and tutorials, supporting inference and fine-tuning of pre-trained models.
  • We have released a vast collection of academic and industrial pretrained models on the ModelScope, which can be accessed through our Model Zoo. The representative Paraformer-large, a non-autoregressive end-to-end speech recognition model, has the advantages of high accuracy, high efficiency, and convenient deployment, supporting the rapid construction of speech recognition services. For more details on service deployment, please refer to the service deployment document.

What's new:

  • 2023/07/17: BAT released a low-latency and low-memory-consumption RNN-T model. For more details, please refer to (BAT).
  • 2023/07/03: The CPU version of the Chinese offline file transcription service has been released with one-click deployment. For more details, please refer to (Deployment documentation).
  • 2023/06/26: ASRU2023 Multi-Channel Multi-Party Meeting Transcription Challenge 2.0 completed the competition and announced the results. For more details, please refer to (M2MeT2.0).

Installation

Please ref to installation docs

Deployment Service

FunASR supports pre-trained or further fine-tuned models for deployment as a service. The CPU version of the Chinese offline file conversion service has been released, details can be found in docs. More detailed information about service deployment can be found in the deployment roadmap.

Quick Start

FunASR supports inference and fine-tuning of models trained on industrial datasets of tens of thousands of hours. For more details, please refer to (modelscope_egs). It also supports training and fine-tuning of models on academic standard datasets. For more details, please refer to(egs). The models include speech recognition (ASR), speech activity detection (VAD), punctuation recovery, language model, speaker verification, speaker separation, and multi-party conversation speech recognition. For a detailed list of models, please refer to the Model Zoo:

Contact

If you have any questions about FunASR, please contact us by

Dingding group Wechat group

Contributors

he contributor list can be found in contributors

License

This project is licensed under the The MIT License. FunASR also contains various third-party components and some code modified from other repos under other open source licenses. The use of pretraining model is subject to model licencs

Citations

@inproceedings{gao2023funasr,
  author={Zhifu Gao and Zerui Li and Jiaming Wang and Haoneng Luo and Xian Shi and Mengzhe Chen and Yabin Li and Lingyun Zuo and Zhihao Du and Zhangyu Xiao and Shiliang Zhang},
  title={FunASR: A Fundamental End-to-End Speech Recognition Toolkit},
  year={2023},
  booktitle={INTERSPEECH},
}
@inproceedings{gao22b_interspeech,
  author={Zhifu Gao and ShiLiang Zhang and Ian McLoughlin and Zhijie Yan},
  title={{Paraformer: Fast and Accurate Parallel Transformer for Non-autoregressive End-to-End Speech Recognition}},
  year=2022,
  booktitle={Proc. Interspeech 2022},
  pages={2063--2067},
  doi={10.21437/Interspeech.2022-9996}
}