| New file |
| | |
| | | /** |
| | | * Copyright FunASR (https://github.com/alibaba-damo-academy/FunASR). All Rights Reserved. |
| | | * MIT License (https://opensource.org/licenses/MIT) |
| | | */ |
| | | |
| | | #include "precomp.h" |
| | | #include "paraformer-torch.h" |
| | | #include "encode_converter.h" |
| | | #include <cstddef> |
| | | |
| | | using namespace std; |
| | | namespace funasr { |
| | | |
| | | ParaformerTorch::ParaformerTorch() |
| | | :use_hotword(false){ |
| | | } |
| | | |
| | | // offline |
| | | void ParaformerTorch::InitAsr(const std::string &am_model, const std::string &am_cmvn, const std::string &am_config, int thread_num){ |
| | | LoadConfigFromYaml(am_config.c_str()); |
| | | // knf options |
| | | fbank_opts_.frame_opts.dither = 0; |
| | | fbank_opts_.mel_opts.num_bins = n_mels; |
| | | fbank_opts_.frame_opts.samp_freq = asr_sample_rate; |
| | | fbank_opts_.frame_opts.window_type = window_type; |
| | | fbank_opts_.frame_opts.frame_shift_ms = frame_shift; |
| | | fbank_opts_.frame_opts.frame_length_ms = frame_length; |
| | | fbank_opts_.energy_floor = 0; |
| | | fbank_opts_.mel_opts.debug_mel = false; |
| | | |
| | | vocab = new Vocab(am_config.c_str()); |
| | | phone_set_ = new PhoneSet(am_config.c_str()); |
| | | LoadCmvn(am_cmvn.c_str()); |
| | | |
| | | torch::DeviceType device = at::kCPU; |
| | | #ifdef USE_GPU |
| | | if (!torch::cuda::is_available()) { |
| | | LOG(ERROR) << "CUDA is not available! Please check your GPU settings"; |
| | | exit(-1); |
| | | } else { |
| | | LOG(INFO) << "CUDA available! Running on GPU"; |
| | | device = at::kCUDA; |
| | | } |
| | | #endif |
| | | #ifdef USE_IPEX |
| | | torch::jit::setTensorExprFuserEnabled(false); |
| | | #endif |
| | | torch::jit::script::Module model = torch::jit::load(am_model, device); |
| | | model_ = std::make_shared<TorchModule>(std::move(model)); |
| | | } |
| | | |
| | | void ParaformerTorch::InitLm(const std::string &lm_file, |
| | | const std::string &lm_cfg_file, |
| | | const std::string &lex_file) { |
| | | try { |
| | | lm_ = std::shared_ptr<fst::Fst<fst::StdArc>>( |
| | | fst::Fst<fst::StdArc>::Read(lm_file)); |
| | | if (lm_){ |
| | | lm_vocab = new Vocab(lm_cfg_file.c_str(), lex_file.c_str()); |
| | | LOG(INFO) << "Successfully load lm file " << lm_file; |
| | | }else{ |
| | | LOG(ERROR) << "Failed to load lm file " << lm_file; |
| | | } |
| | | } catch (std::exception const &e) { |
| | | LOG(ERROR) << "Error when load lm file: " << e.what(); |
| | | exit(0); |
| | | } |
| | | } |
| | | |
| | | void ParaformerTorch::LoadConfigFromYaml(const char* filename){ |
| | | |
| | | YAML::Node config; |
| | | try{ |
| | | config = YAML::LoadFile(filename); |
| | | }catch(exception const &e){ |
| | | LOG(ERROR) << "Error loading file, yaml file error or not exist."; |
| | | exit(-1); |
| | | } |
| | | |
| | | try{ |
| | | YAML::Node frontend_conf = config["frontend_conf"]; |
| | | this->asr_sample_rate = frontend_conf["fs"].as<int>(); |
| | | |
| | | YAML::Node lang_conf = config["lang"]; |
| | | if (lang_conf.IsDefined()){ |
| | | language = lang_conf.as<string>(); |
| | | } |
| | | }catch(exception const &e){ |
| | | LOG(ERROR) << "Error when load argument from vad config YAML."; |
| | | exit(-1); |
| | | } |
| | | } |
| | | |
| | | void ParaformerTorch::InitHwCompiler(const std::string &hw_model, int thread_num) { |
| | | // TODO |
| | | use_hotword = true; |
| | | } |
| | | |
| | | void ParaformerTorch::InitSegDict(const std::string &seg_dict_model) { |
| | | seg_dict = new SegDict(seg_dict_model.c_str()); |
| | | } |
| | | |
| | | ParaformerTorch::~ParaformerTorch() |
| | | { |
| | | if(vocab){ |
| | | delete vocab; |
| | | } |
| | | if(lm_vocab){ |
| | | delete lm_vocab; |
| | | } |
| | | if(seg_dict){ |
| | | delete seg_dict; |
| | | } |
| | | if(phone_set_){ |
| | | delete phone_set_; |
| | | } |
| | | } |
| | | |
| | | void ParaformerTorch::StartUtterance() |
| | | { |
| | | } |
| | | |
| | | void ParaformerTorch::EndUtterance() |
| | | { |
| | | } |
| | | |
| | | void ParaformerTorch::Reset() |
| | | { |
| | | } |
| | | |
| | | void ParaformerTorch::FbankKaldi(float sample_rate, const float* waves, int len, std::vector<std::vector<float>> &asr_feats) { |
| | | knf::OnlineFbank fbank_(fbank_opts_); |
| | | std::vector<float> buf(len); |
| | | for (int32_t i = 0; i != len; ++i) { |
| | | buf[i] = waves[i] * 32768; |
| | | } |
| | | fbank_.AcceptWaveform(sample_rate, buf.data(), buf.size()); |
| | | |
| | | int32_t frames = fbank_.NumFramesReady(); |
| | | for (int32_t i = 0; i != frames; ++i) { |
| | | const float *frame = fbank_.GetFrame(i); |
| | | std::vector<float> frame_vector(frame, frame + fbank_opts_.mel_opts.num_bins); |
| | | asr_feats.emplace_back(frame_vector); |
| | | } |
| | | } |
| | | |
| | | void ParaformerTorch::LoadCmvn(const char *filename) |
| | | { |
| | | ifstream cmvn_stream(filename); |
| | | if (!cmvn_stream.is_open()) { |
| | | LOG(ERROR) << "Failed to open file: " << filename; |
| | | exit(-1); |
| | | } |
| | | string line; |
| | | |
| | | while (getline(cmvn_stream, line)) { |
| | | istringstream iss(line); |
| | | vector<string> line_item{istream_iterator<string>{iss}, istream_iterator<string>{}}; |
| | | if (line_item[0] == "<AddShift>") { |
| | | getline(cmvn_stream, line); |
| | | istringstream means_lines_stream(line); |
| | | vector<string> means_lines{istream_iterator<string>{means_lines_stream}, istream_iterator<string>{}}; |
| | | if (means_lines[0] == "<LearnRateCoef>") { |
| | | for (int j = 3; j < means_lines.size() - 1; j++) { |
| | | means_list_.push_back(stof(means_lines[j])); |
| | | } |
| | | continue; |
| | | } |
| | | } |
| | | else if (line_item[0] == "<Rescale>") { |
| | | getline(cmvn_stream, line); |
| | | istringstream vars_lines_stream(line); |
| | | vector<string> vars_lines{istream_iterator<string>{vars_lines_stream}, istream_iterator<string>{}}; |
| | | if (vars_lines[0] == "<LearnRateCoef>") { |
| | | for (int j = 3; j < vars_lines.size() - 1; j++) { |
| | | vars_list_.push_back(stof(vars_lines[j])*scale); |
| | | } |
| | | continue; |
| | | } |
| | | } |
| | | } |
| | | } |
| | | |
| | | string ParaformerTorch::GreedySearch(float * in, int n_len, int64_t token_nums, bool is_stamp, std::vector<float> us_alphas, std::vector<float> us_cif_peak) |
| | | { |
| | | vector<int> hyps; |
| | | int Tmax = n_len; |
| | | for (int i = 0; i < Tmax; i++) { |
| | | int max_idx; |
| | | float max_val; |
| | | FindMax(in + i * token_nums, token_nums, max_val, max_idx); |
| | | hyps.push_back(max_idx); |
| | | } |
| | | if(!is_stamp){ |
| | | return vocab->Vector2StringV2(hyps, language); |
| | | }else{ |
| | | std::vector<string> char_list; |
| | | std::vector<std::vector<float>> timestamp_list; |
| | | std::string res_str; |
| | | vocab->Vector2String(hyps, char_list); |
| | | std::vector<string> raw_char(char_list); |
| | | TimestampOnnx(us_alphas, us_cif_peak, char_list, res_str, timestamp_list); |
| | | |
| | | return PostProcess(raw_char, timestamp_list); |
| | | } |
| | | } |
| | | |
| | | string ParaformerTorch::BeamSearch(WfstDecoder* &wfst_decoder, float *in, int len, int64_t token_nums) |
| | | { |
| | | return wfst_decoder->Search(in, len, token_nums); |
| | | } |
| | | |
| | | string ParaformerTorch::FinalizeDecode(WfstDecoder* &wfst_decoder, |
| | | bool is_stamp, std::vector<float> us_alphas, std::vector<float> us_cif_peak) |
| | | { |
| | | return wfst_decoder->FinalizeDecode(is_stamp, us_alphas, us_cif_peak); |
| | | } |
| | | |
| | | void ParaformerTorch::LfrCmvn(std::vector<std::vector<float>> &asr_feats) { |
| | | |
| | | std::vector<std::vector<float>> out_feats; |
| | | int T = asr_feats.size(); |
| | | int T_lrf = ceil(1.0 * T / lfr_n); |
| | | |
| | | // Pad frames at start(copy first frame) |
| | | for (int i = 0; i < (lfr_m - 1) / 2; i++) { |
| | | asr_feats.insert(asr_feats.begin(), asr_feats[0]); |
| | | } |
| | | // Merge lfr_m frames as one,lfr_n frames per window |
| | | T = T + (lfr_m - 1) / 2; |
| | | std::vector<float> p; |
| | | for (int i = 0; i < T_lrf; i++) { |
| | | if (lfr_m <= T - i * lfr_n) { |
| | | for (int j = 0; j < lfr_m; j++) { |
| | | p.insert(p.end(), asr_feats[i * lfr_n + j].begin(), asr_feats[i * lfr_n + j].end()); |
| | | } |
| | | out_feats.emplace_back(p); |
| | | p.clear(); |
| | | } else { |
| | | // Fill to lfr_m frames at last window if less than lfr_m frames (copy last frame) |
| | | int num_padding = lfr_m - (T - i * lfr_n); |
| | | for (int j = 0; j < (asr_feats.size() - i * lfr_n); j++) { |
| | | p.insert(p.end(), asr_feats[i * lfr_n + j].begin(), asr_feats[i * lfr_n + j].end()); |
| | | } |
| | | for (int j = 0; j < num_padding; j++) { |
| | | p.insert(p.end(), asr_feats[asr_feats.size() - 1].begin(), asr_feats[asr_feats.size() - 1].end()); |
| | | } |
| | | out_feats.emplace_back(p); |
| | | p.clear(); |
| | | } |
| | | } |
| | | // Apply cmvn |
| | | for (auto &out_feat: out_feats) { |
| | | for (int j = 0; j < means_list_.size(); j++) { |
| | | out_feat[j] = (out_feat[j] + means_list_[j]) * vars_list_[j]; |
| | | } |
| | | } |
| | | asr_feats = out_feats; |
| | | } |
| | | |
| | | string ParaformerTorch::Forward(float* din, int len, bool input_finished, const std::vector<std::vector<float>> &hw_emb, void* decoder_handle) |
| | | { |
| | | WfstDecoder* wfst_decoder = (WfstDecoder*)decoder_handle; |
| | | int32_t in_feat_dim = fbank_opts_.mel_opts.num_bins; |
| | | |
| | | std::vector<std::vector<float>> asr_feats; |
| | | FbankKaldi(asr_sample_rate, din, len, asr_feats); |
| | | if(asr_feats.size() == 0){ |
| | | return ""; |
| | | } |
| | | LfrCmvn(asr_feats); |
| | | int32_t feat_dim = lfr_m*in_feat_dim; |
| | | int32_t num_frames = asr_feats.size(); |
| | | |
| | | std::vector<float> wav_feats; |
| | | for (const auto &frame_feat: asr_feats) { |
| | | wav_feats.insert(wav_feats.end(), frame_feat.begin(), frame_feat.end()); |
| | | } |
| | | std::vector<int32_t> paraformer_length; |
| | | paraformer_length.emplace_back(num_frames); |
| | | |
| | | torch::NoGradGuard no_grad; |
| | | torch::Tensor feats = |
| | | torch::from_blob(wav_feats.data(), |
| | | {1, num_frames, feat_dim}, torch::kFloat).contiguous(); |
| | | torch::Tensor feat_lens = torch::from_blob(paraformer_length.data(), |
| | | {1}, torch::kInt32); |
| | | |
| | | // 2. forward |
| | | #ifdef USE_GPU |
| | | feats = feats.to(at::kCUDA); |
| | | feat_lens = feat_lens.to(at::kCUDA); |
| | | #endif |
| | | std::vector<torch::jit::IValue> inputs = {feats, feat_lens}; |
| | | |
| | | string result=""; |
| | | try { |
| | | auto outputs = model_->forward(inputs).toTuple()->elements(); |
| | | torch::Tensor am_scores; |
| | | torch::Tensor valid_token_lens; |
| | | #ifdef USE_GPU |
| | | am_scores = outputs[0].toTensor().to(at::kCPU); |
| | | valid_token_lens = outputs[1].toTensor().to(at::kCPU); |
| | | #else |
| | | am_scores = outputs[0].toTensor(); |
| | | valid_token_lens = outputs[1].toTensor(); |
| | | #endif |
| | | |
| | | if (lm_ == nullptr) { |
| | | result = GreedySearch(am_scores[0].data_ptr<float>(), valid_token_lens[0].item<int>(), am_scores.size(2)); |
| | | } else { |
| | | result = BeamSearch(wfst_decoder, am_scores[0].data_ptr<float>(), valid_token_lens[0].item<int>(), am_scores.size(2)); |
| | | if (input_finished) { |
| | | result = FinalizeDecode(wfst_decoder); |
| | | } |
| | | } |
| | | } |
| | | catch (std::exception const &e) |
| | | { |
| | | LOG(ERROR)<<e.what(); |
| | | } |
| | | |
| | | return result; |
| | | } |
| | | |
| | | std::vector<std::vector<float>> ParaformerTorch::CompileHotwordEmbedding(std::string &hotwords) { |
| | | std::vector<std::vector<float>> result; |
| | | return result; |
| | | } |
| | | |
| | | Vocab* ParaformerTorch::GetVocab() |
| | | { |
| | | return vocab; |
| | | } |
| | | |
| | | Vocab* ParaformerTorch::GetLmVocab() |
| | | { |
| | | return lm_vocab; |
| | | } |
| | | |
| | | PhoneSet* ParaformerTorch::GetPhoneSet() |
| | | { |
| | | return phone_set_; |
| | | } |
| | | |
| | | string ParaformerTorch::Rescoring() |
| | | { |
| | | LOG(ERROR)<<"Not Imp!!!!!!"; |
| | | return ""; |
| | | } |
| | | } // namespace funasr |