1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
| #!/usr/bin/env python3
| # -*- encoding: utf-8 -*-
| # Copyright FunASR (https://github.com/alibaba-damo-academy/FunASR). All Rights Reserved.
| # MIT License (https://opensource.org/licenses/MIT)
|
| import json
| import os
| import sys
|
| from funasr import AutoModel
|
| ckpt_dir = "/nfs/beinian.lzr/workspace/GPT-4o/Exp/exp6/5m-8gpu/exp6_speech2text_linear_ddp_0609"
| ckpt_id = "model.pt.ep0.90000"
| jsonl = (
| "/nfs/beinian.lzr/workspace/GPT-4o/Data/Speech2Text/TestData/aishell1_test_speech2text.jsonl"
| )
| output_dir = f"{os.path.join(ckpt_dir, ckpt_id)}"
| device = "cuda:0"
|
| ckpt_dir = sys.argv[1]
| ckpt_id = sys.argv[2]
| jsonl = sys.argv[3]
| output_dir = sys.argv[4]
| device = sys.argv[5]
|
| model = AutoModel(
| model=ckpt_dir,
| init_param=f"{os.path.join(ckpt_dir, ckpt_id)}",
| output_dir=output_dir,
| device=device,
| fp16=False,
| bf16=False,
| llm_dtype="bf16",
| )
|
|
| with open(jsonl, "r") as f:
| lines = f.readlines()
|
| tearchforing = False
| for i, line in enumerate(lines):
| data_dict = json.loads(line.strip())
| data = data_dict["messages"]
|
| res = model.generate(
| input=[data],
| tearchforing=tearchforing,
| cache={},
| )
|
| print(res)
|
|