游雁 2 лет назад
Родитель
Сommit
99340740f5
2 измененных файлов с 2 добавлено и 2 удалено
  1. 1 1
      funasr/cli/train_cli.py
  2. 1 1
      funasr/cli/trainer.py

+ 1 - 1
funasr/cli/train_cli.py

@@ -46,7 +46,7 @@ def main(kwargs: DictConfig):
 	
 	local_rank = int(os.environ.get('LOCAL_RANK', 0))
 	# Check if we are using DDP or FSDP
-	use_ddp = 'WORLD_SIZE' in os.environ and os.environ["WORLD_SIZE"] > 1
+	use_ddp = 'WORLD_SIZE' in os.environ and int(os.environ["WORLD_SIZE"]) > 1
 	use_fsdp = kwargs.get("use_fsdp", None)
 	if use_ddp or use_fsdp:
 		dist.init_process_group(backend=kwargs.get("backend", "nccl"), init_method='env://')

+ 1 - 1
funasr/cli/trainer.py

@@ -180,7 +180,7 @@ class Trainer:
 			pbar.update(1)
 			if self.local_rank == 0:
 				pbar.set_description(
-					f"Training Epoch: {epoch + 1}/{self.max_epoch}, step {batch_idx}/{len(self.dataloader_train)}  (loss: {loss.detach().float()})")
+					f"Training Epoch: {epoch + 1}/{self.max_epoch}, step {batch_idx}/{len(self.dataloader_train)}  (loss: {loss.detach().float():.3f}, {[(k, round(v.cpu().item(), 3)) for k, v in stats.items()]})")
 			
 		pbar.close()