| | |
| | | import torch |
| | | from torch.nn import functional as F |
| | | from funasr.models.encoder.abs_encoder import AbsEncoder |
| | | from typing import Tuple, Optional |
| | | from funasr.models.pooling.statistic_pooling import statistic_pooling, windowed_statistic_pooling |
| | | from funasr.models.encoder.abs_encoder import AbsEncoder |
| | | from collections import OrderedDict |
| | | import logging |
| | | import numpy as np |
| | |
| | | return xs_pad, ilens |
| | | |
| | | |
| | | class ResNet34(AbsEncoder): |
| | | class ResNet34(torch.nn.Module): |
| | | def __init__( |
| | | self, |
| | | input_size, |
| | |
| | | "{}.resnet{}_dense.weight".format(tensor_name_prefix_torch, layer_idx): |
| | | {"name": "{}/resnet{}_dense/kernel".format(tensor_name_prefix_tf, layer_idx), |
| | | "squeeze": None, |
| | | "transpose": (3, 2, 0, 1) if layer_idx == 0 else (1, 0), |
| | | "transpose": (2, 1, 0) if layer_idx == 0 else (1, 0), |
| | | }, |
| | | "{}.resnet{}_dense.bias".format(tensor_name_prefix_torch, layer_idx): |
| | | {"name": "{}/resnet{}_dense/bias".format(tensor_name_prefix_tf, layer_idx), |
| | |
| | | name, data_tf.size(), name_tf, var_dict_tf[name_tf].shape |
| | | )) |
| | | else: |
| | | var_dict_torch_update[name] = torch.Tensor(map_dict[name]).type(torch.int64).to("cpu") |
| | | var_dict_torch_update[name] = torch.from_numpy(np.array(map_dict[name])).type(torch.int64).to("cpu") |
| | | logging.info("torch tensor: {}, manually assigning to: {}".format( |
| | | name, map_dict[name] |
| | | )) |