雾聪
2024-03-28 7963f259ea19cbab7970fa9ec5ac7e852a5d9799
runtime/onnxruntime/src/funasrruntime.cpp
@@ -33,9 +33,9 @@
      return mm;
   }
   _FUNASRAPI FUNASR_HANDLE  FunOfflineInit(std::map<std::string, std::string>& model_path, int thread_num)
   _FUNASRAPI FUNASR_HANDLE  FunOfflineInit(std::map<std::string, std::string>& model_path, int thread_num, bool use_gpu)
   {
      funasr::OfflineStream* mm = funasr::CreateOfflineStream(model_path, thread_num);
      funasr::OfflineStream* mm = funasr::CreateOfflineStream(model_path, thread_num, use_gpu);
      return mm;
   }
@@ -146,6 +146,7 @@
      funasr::FUNASR_VAD_RESULT* p_result = new funasr::FUNASR_VAD_RESULT;
      p_result->snippet_time = audio.GetTimeLen();
      if(p_result->snippet_time == 0){
         p_result->segments = new vector<std::vector<int>>();
            return p_result;
        }
      
@@ -178,6 +179,7 @@
      funasr::FUNASR_VAD_RESULT* p_result = new funasr::FUNASR_VAD_RESULT;
      p_result->snippet_time = audio.GetTimeLen();
      if(p_result->snippet_time == 0){
         p_result->segments = new vector<std::vector<int>>();
            return p_result;
        }
      
@@ -243,15 +245,13 @@
            return p_result;
        }
      if(offline_stream->UseVad()){
         audio.Split(offline_stream);
         audio.CutSplit(offline_stream);
      }
      float* buff;
      int len;
      int flag = 0;
      int n_step = 0;
      int n_total = audio.GetQueueSize();
      float start_time = 0.0;
      std::string cur_stamp = "[";
      std::string lang = (offline_stream->asr_handle)->GetLang();
@@ -279,9 +279,6 @@
               cur_stamp += "["+std::to_string((int)(1000*begin))+","+std::to_string((int)(1000*end))+"],";
            }
         }
         n_step++;
         if (fn_callback)
            fn_callback(n_step, n_total);
      }
      if(cur_stamp != "["){
         cur_stamp.erase(cur_stamp.length() - 1);
@@ -303,7 +300,9 @@
         p_result->msg = msg_itn;
      }
#endif
      if (!(p_result->stamp).empty()){
         p_result->stamp_sents = funasr::TimestampSentence(p_result->msg, p_result->stamp);
      }
      return p_result;
   }
@@ -339,7 +338,7 @@
            return p_result;
        }
      if(offline_stream->UseVad()){
         audio.Split(offline_stream);
         audio.CutSplit(offline_stream);
      }
      float* buff;
@@ -399,6 +398,9 @@
         p_result->msg = msg_itn;
      }
#endif
      if (!(p_result->stamp).empty()){
         p_result->stamp_sents = funasr::TimestampSentence(p_result->msg, p_result->stamp);
      }
      return p_result;
   }
@@ -432,7 +434,7 @@
   _FUNASRAPI FUNASR_RESULT FunTpassInferBuffer(FUNASR_HANDLE handle, FUNASR_HANDLE online_handle, const char* sz_buf, 
                                     int n_len, std::vector<std::vector<std::string>> &punc_cache, bool input_finished, 
                                     int sampling_rate, std::string wav_format, ASR_TYPE mode, 
                                     const std::vector<std::vector<float>> &hw_emb, bool itn)
                                     const std::vector<std::vector<float>> &hw_emb, bool itn, FUNASR_DEC_HANDLE dec_handle)
   {
      funasr::TpassStream* tpass_stream = (funasr::TpassStream*)handle;
      funasr::TpassOnlineStream* tpass_online_stream = (funasr::TpassOnlineStream*)online_handle;
@@ -473,7 +475,7 @@
      
      audio->Split(vad_online_handle, chunk_len, input_finished, mode);
      funasr::AudioFrame* frame = NULL;
      funasr::AudioFrame* frame = nullptr;
      while(audio->FetchChunck(frame) > 0){
         string msg = ((funasr::ParaformerOnline*)asr_online_handle)->Forward(frame->data, frame->len, frame->is_final);
         if(mode == ASR_ONLINE){
@@ -497,16 +499,21 @@
         }else if(mode == ASR_TWO_PASS){
            p_result->msg += msg;
         }
         if(frame != NULL){
         if(frame != nullptr){
            delete frame;
            frame = NULL;
            frame = nullptr;
         }
      }
      // timestamp
      std::string cur_stamp = "[";      
      while(audio->FetchTpass(frame) > 0){
         string msg = ((funasr::Paraformer*)asr_handle)->Forward(frame->data, frame->len, frame->is_final, hw_emb);
         // dec reset
         funasr::WfstDecoder* wfst_decoder = (funasr::WfstDecoder*)dec_handle;
         if (wfst_decoder){
            wfst_decoder->StartUtterance();
         }
         string msg = ((funasr::Paraformer*)asr_handle)->Forward(frame->data, frame->len, frame->is_final, hw_emb, dec_handle);
         std::vector<std::string> msg_vec = funasr::split(msg, '|');  // split with timestamp
         if(msg_vec.size()==0){
@@ -546,10 +553,12 @@
            p_result->tpass_msg = msg_itn;
         }
#endif
         if(frame != NULL){
         if (!(p_result->stamp).empty()){
            p_result->stamp_sents = funasr::TimestampSentence(p_result->tpass_msg, p_result->stamp);
         }
         if(frame != nullptr){
            delete frame;
            frame = NULL;
            frame = nullptr;
         }
      }
@@ -601,6 +610,15 @@
         return nullptr;
      return p_result->stamp.c_str();
   }
      _FUNASRAPI const char* FunASRGetStampSents(FUNASR_RESULT result)
   {
      funasr::FUNASR_RECOG_RESULT * p_result = (funasr::FUNASR_RECOG_RESULT*)result;
      if(!p_result)
         return nullptr;
      return p_result->stamp_sents.c_str();
   }
   _FUNASRAPI const char* FunASRGetTpassResult(FUNASR_RESULT result,int n_index)
@@ -744,10 +762,45 @@
      funasr::WfstDecoder* mm = nullptr;
      if (asr_type == ASR_OFFLINE) {
         funasr::OfflineStream* offline_stream = (funasr::OfflineStream*)handle;
         funasr::Paraformer* paraformer = (funasr::Paraformer*)offline_stream->asr_handle.get();
         if (paraformer->lm_)
            mm = new funasr::WfstDecoder(paraformer->lm_.get(),
               paraformer->GetPhoneSet(), paraformer->GetVocab(), glob_beam, lat_beam, am_scale);
         auto paraformer = dynamic_cast<funasr::Paraformer*>(offline_stream->asr_handle.get());
         if(paraformer !=nullptr){
            if (paraformer->lm_){
               mm = new funasr::WfstDecoder(paraformer->lm_.get(),
                  paraformer->GetPhoneSet(), paraformer->GetLmVocab(), glob_beam, lat_beam, am_scale);
            }
            return mm;
         }
         #ifdef USE_GPU
         auto paraformer_torch = dynamic_cast<funasr::ParaformerTorch*>(offline_stream->asr_handle.get());
         if(paraformer_torch !=nullptr){
            if (paraformer_torch->lm_){
               mm = new funasr::WfstDecoder(paraformer_torch->lm_.get(),
                  paraformer_torch->GetPhoneSet(), paraformer_torch->GetLmVocab(), glob_beam, lat_beam, am_scale);
            }
            return mm;
         }
         #endif
      } else if (asr_type == ASR_TWO_PASS){
         funasr::TpassStream* tpass_stream = (funasr::TpassStream*)handle;
         auto paraformer = dynamic_cast<funasr::Paraformer*>(tpass_stream->asr_handle.get());
         if(paraformer !=nullptr){
            if (paraformer->lm_){
               mm = new funasr::WfstDecoder(paraformer->lm_.get(),
                  paraformer->GetPhoneSet(), paraformer->GetLmVocab(), glob_beam, lat_beam, am_scale);
            }
            return mm;
         }
         #ifdef USE_GPU
         auto paraformer_torch = dynamic_cast<funasr::ParaformerTorch*>(tpass_stream->asr_handle.get());
         if(paraformer_torch !=nullptr){
            if (paraformer_torch->lm_){
               mm = new funasr::WfstDecoder(paraformer_torch->lm_.get(),
                  paraformer_torch->GetPhoneSet(), paraformer_torch->GetLmVocab(), glob_beam, lat_beam, am_scale);
            }
            return mm;
         }
         #endif
      }
      return mm;
   }