Paraformer
pretrained model in ModelScope:speech_paraformer_asr_nat-aishell1-pytorch
Training Config
- Feature info: using 80 dims fbank, global cmvn, speed perturb(0.9, 1.0, 1.1), specaugment
- Train info: lr 5e-4, batch_size 25000, 2 gpu(Tesla V100), acc_grad 1, 50 epochs
- Train config: conf/train_asr_paraformer_conformer_12e_6d_2048_256.yaml
- LM config: LM was not used
Results (CER)
- Decode config: conf/decode_asr_transformer_noctc_1best.yaml (ctc weight:0.0)
| testset |
CER(%) |
| dev |
4.66 |
| test |
5.11 |
- Decode config: conf/decode_asr_transformer.yaml (ctc weight:0.5)
| testset |
CER(%) |
| dev |
4.52 |
| test |
4.94 |