游雁 1 жил өмнө
parent
commit
26ab38b56c

+ 1 - 0
examples/aishell/branchformer/run.sh

@@ -174,6 +174,7 @@ if [ ${stage} -le 5 ] && [ ${stop_stage} -ge 5 ]; then
           ++output_dir="${inference_dir}/${JOB}" \
           ++device="${inference_device}" \
           ++ncpu=1 \
+          ++disable_log=true \
           ++batch_size="${inference_batch_size}" &> ${_logdir}/log.${JOB}.txt
         }&
 

+ 1 - 0
examples/aishell/conformer/run.sh

@@ -173,6 +173,7 @@ if [ ${stage} -le 5 ] && [ ${stop_stage} -ge 5 ]; then
           ++output_dir="${inference_dir}/${JOB}" \
           ++device="${inference_device}" \
           ++ncpu=1 \
+          ++disable_log=true \
           ++batch_size="${inference_batch_size}" &> ${_logdir}/log.${JOB}.txt
         }&
 

+ 1 - 0
examples/aishell/e_branchformer/run.sh

@@ -174,6 +174,7 @@ if [ ${stage} -le 5 ] && [ ${stop_stage} -ge 5 ]; then
           ++output_dir="${inference_dir}/${JOB}" \
           ++device="${inference_device}" \
           ++ncpu=1 \
+          ++disable_log=true \
           ++batch_size="${inference_batch_size}" &> ${_logdir}/log.${JOB}.txt
         }&
 

+ 1 - 0
examples/aishell/paraformer/run.sh

@@ -173,6 +173,7 @@ if [ ${stage} -le 5 ] && [ ${stop_stage} -ge 5 ]; then
           ++output_dir="${inference_dir}/${JOB}" \
           ++device="${inference_device}" \
           ++ncpu=1 \
+          ++disable_log=true \
           ++batch_size="${inference_batch_size}" &> ${_logdir}/log.${JOB}.txt
         }&
 

+ 1 - 0
examples/aishell/transformer/run.sh

@@ -174,6 +174,7 @@ if [ ${stage} -le 5 ] && [ ${stop_stage} -ge 5 ]; then
           ++output_dir="${inference_dir}/${JOB}" \
           ++device="${inference_device}" \
           ++ncpu=1 \
+          ++disable_log=true \
           ++batch_size="${inference_batch_size}" &> ${_logdir}/log.${JOB}.txt
         }&
 

+ 4 - 0
funasr/train_utils/trainer.py

@@ -3,6 +3,7 @@ import time
 import torch
 import logging
 from tqdm import tqdm
+from datetime import datetime
 import torch.distributed as dist
 from contextlib import nullcontext
 # from torch.utils.tensorboard import SummaryWriter
@@ -283,7 +284,10 @@ class Trainer:
                                              torch.cuda.max_memory_reserved()/1024/1024/1024,
                                              )
                 lr = self.scheduler.get_last_lr()[0]
+                time_now = datetime.now()
+                time_now = now.strftime("%Y-%m-%d %H:%M:%S")
                 description = (
+                    f"{time_now}, "
                     f"rank: {self.local_rank}, "
                     f"epoch: {epoch}/{self.max_epoch}, "
                     f"step: {batch_idx+1}/{len(self.dataloader_train)}, total: {self.batch_total}, "