| | |
| | | filelist = [".scp", ".txt", ".json", ".jsonl", ".text"] |
| | | |
| | | chars = string.ascii_letters + string.digits |
| | | if isinstance(data_in, str) and data_in.startswith("http"): # url |
| | | data_in = download_from_url(data_in) |
| | | if isinstance(data_in, str): |
| | | if data_in.startswith("http://") or data_in.startswith("https://"): # url |
| | | data_in = download_from_url(data_in) |
| | | |
| | | if isinstance(data_in, str) and os.path.exists( |
| | | data_in |
| | |
| | | # fp16 |
| | | if kwargs.get("fp16", False): |
| | | model.to(torch.float16) |
| | | elif kwargs.get("bf16", False): |
| | | model.to(torch.bfloat16) |
| | | return model, kwargs |
| | | |
| | | def __call__(self, *args, **cfg): |
| | |
| | | with torch.no_grad(): |
| | | res = model.inference(**batch, **kwargs) |
| | | if isinstance(res, (list, tuple)): |
| | | results = res[0] |
| | | results = res[0] if len(res) > 0 else [{"text": ""}] |
| | | meta_data = res[1] if len(res) > 1 else {} |
| | | time2 = time.perf_counter() |
| | | |
| | |
| | | results_sorted = [] |
| | | |
| | | if not len(sorted_data): |
| | | results_ret_list.append({"key": key, "text": "", "timestamp": []}) |
| | | logging.info("decoding, utt: {}, empty speech".format(key)) |
| | | continue |
| | | |
| | |
| | | # f"time_speech_total_per_sample: {time_speech_total_per_sample: 0.3f}, " |
| | | # f"time_escape_total_per_sample: {time_escape_total_per_sample:0.3f}") |
| | | |
| | | if len(results_sorted) != n: |
| | | results_ret_list.append({"key": key, "text": "", "timestamp": []}) |
| | | logging.info("decoding, utt: {}, empty result".format(key)) |
| | | continue |
| | | restored_data = [0] * n |
| | | for j in range(n): |
| | | index = sorted_data[j][1] |