From 4f7234fd821e623b3eba1cec2f2ee8d23609ec64 Mon Sep 17 00:00:00 2001
From: huangmingming <huangmingming@deepscience.cn>
Date: 星期一, 30 一月 2023 16:19:12 +0800
Subject: [PATCH] update tips

---
 funasr/runtime/python/grpc/grpc_main_client_mic.py |   26 ++++++++------------------
 1 files changed, 8 insertions(+), 18 deletions(-)

diff --git a/funasr/runtime/python/grpc/grpc_main_client_mic.py b/funasr/runtime/python/grpc/grpc_main_client_mic.py
index de2cded..8a8fe4d 100644
--- a/funasr/runtime/python/grpc/grpc_main_client_mic.py
+++ b/funasr/runtime/python/grpc/grpc_main_client_mic.py
@@ -10,21 +10,14 @@
 import time
 import asyncio
 import datetime
-
-SPEAKING = False
-stub = None
-asr_user = None
-language = None
+import argparse
 
 async def deal_chunk(sig_mic):
-    
-    global stub,SPEAKING,asr_user,language
+    global stub,SPEAKING,asr_user,language,sample_rate
     sig = np.frombuffer(sig_mic, 'int16')
     if vad.is_speech(sig.tobytes(), sample_rate): #speaking
         SPEAKING = True
         response = transcribe_audio_bytes(stub, sig, user=asr_user, language=language, speaking = True, isEnd = False) #speaking, send audio to server.
-        #print("response")
-        #print (response.next())
     else: #silence   
         begin_time = 0
         if SPEAKING: #means we have some audio recorded, send recognize order to server.
@@ -33,17 +26,12 @@
             response = transcribe_audio_bytes(stub, None, user=asr_user, language=language, speaking = False, isEnd = False) #speak end, call server for recognize one sentence
             resp = response.next()           
             if "decoding" == resp.action:   
-                print(resp.action)
-                print(json.loads(resp.sentence))
                 resp = response.next() #TODO, blocking operation may leads to miss some audio clips. C++ multi-threading is preferred.
                 if "finish" == resp.action:        
                     end_time = int(round(time.time() * 1000))
                     print (json.loads(resp.sentence))
-                    #print ("silence, end_time: %d " % end_time)
                     print ("delay in ms: %d " % (end_time - begin_time))
                 else:
-                    #debug
-                    print (resp.action + " " + str(json.loads(resp.sentence)))
                     pass
         
 
@@ -58,7 +46,7 @@
 
         #end grpc
         response = transcribe_audio_bytes(stub, None, user=asr_user, language=language, speaking = False, isEnd = True)
-        #print (response.next())
+        print (response.next().action)
 
 
 if __name__ == '__main__':
@@ -97,10 +85,12 @@
 
     args = parser.parse_args()
     
-    global SPEAKING,asr_user,language
+
     SPEAKING = False
-    asr_user = args.asr_user
+    asr_user = args.user_allowed
+    sample_rate = args.sample_rate
     language = 'zh-CN'  
+    
 
     vad = webrtcvad.Vad()
     vad.set_mode(1)
@@ -116,7 +106,7 @@
                 frames_per_buffer=args.mic_chunk)
                 
     print("* recording")
-    asyncio.run(record(args.host,args.port,args.sample_rate,args.mic_chunk,args.record_seconds,args.asr_user,args.language))
+    asyncio.run(record(args.host,args.port,args.sample_rate,args.mic_chunk,args.record_seconds,args.user_allowed,language))
     stream.stop_stream()
     stream.close()
     p.terminate()

--
Gitblit v1.9.1