Browse Source

update server env

huangmingming 3 years ago
parent
commit
ddce9cc877
1 changed files with 5 additions and 5 deletions
  1. 5 5
      funasr/runtime/python/grpc/Readme.md

+ 5 - 5
funasr/runtime/python/grpc/Readme.md

@@ -7,13 +7,13 @@ The audio data is in streaming, the asr inference process is in offline.
 
 Step 1) Prepare server environment (on server).
 ```
-# Optional, modelscope cuda docker is preferred.
-CID=`docker run --network host -d -it --gpus '"device=0"' registry.cn-hangzhou.aliyuncs.com/modelscope-repo/modelscope:ubuntu20.04-cuda11.3.0-py37-torch1.11.0-tf1.15.5-1.2.0`
-echo $CID
-docker exec -it $CID /bin/bash
+# Install modelscope and funasr, or install with modelscope cuda-docker image. 
+
+# Get into grpc directory.
 cd /opt/conda/lib/python3.7/site-packages/funasr/runtime/python/grpc
 ```
 
+
 Step 2) Generate protobuf file (for server and client).
 ```
 # Optional, paraformer_pb2.py and paraformer_pb2_grpc.py are already generated.
@@ -41,7 +41,7 @@ python grpc_main_client_mic.py --host 127.0.0.1 --port 10095
 
 
 ## Reference
-We borrow or refer to some code from:
+We borrow from or refer to some code as:
 
 1)https://github.com/wenet-e2e/wenet/tree/main/runtime/core/grpc